Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
35,944
| 14,901,293,763
|
IssuesEvent
|
2021-01-21 16:18:34
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
Walk-through MS Form for SMB Career Development Survey
|
Service: Apps Type: Data Workgroup: SMB
|
Susanne was working on a form in Survey123 and I found out about it. I reached out to her and told her about MS Forms. I volunteered to convert her form because she had to be in an RFP meeting all afternoon.
|
1.0
|
Walk-through MS Form for SMB Career Development Survey - Susanne was working on a form in Survey123 and I found out about it. I reached out to her and told her about MS Forms. I volunteered to convert her form because she had to be in an RFP meeting all afternoon.
|
non_test
|
walk through ms form for smb career development survey susanne was working on a form in and i found out about it i reached out to her and told her about ms forms i volunteered to convert her form because she had to be in an rfp meeting all afternoon
| 0
|
181,247
| 21,657,939,731
|
IssuesEvent
|
2022-05-06 15:53:19
|
turkdevops/apps
|
https://api.github.com/repos/turkdevops/apps
|
closed
|
CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - autoclosed
|
security vulnerability
|
## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/mocha/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- mocha-5.2.0.tgz (Root Library)
- mkdirp-0.5.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/meow/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- imagemin-pngquant-8.0.0.tgz (Root Library)
- pngquant-bin-5.0.2.tgz
- logalot-2.1.0.tgz
- squeak-1.3.0.tgz
- lpad-align-1.1.2.tgz
- meow-3.7.0.tgz
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/apps/commit/6b6da833aa3bd0bc4a04b5931212e0ecf31d7235">6b6da833aa3bd0bc4a04b5931212e0ecf31d7235</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution (minimist): 0.2.1</p>
<p>Direct dependency fix Resolution (mocha): 6.2.3</p><p>Fix Resolution (minimist): 1.2.3</p>
<p>Direct dependency fix Resolution (imagemin-pngquant): 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - autoclosed - ## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/mocha/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- mocha-5.2.0.tgz (Root Library)
- mkdirp-0.5.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/meow/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- imagemin-pngquant-8.0.0.tgz (Root Library)
- pngquant-bin-5.0.2.tgz
- logalot-2.1.0.tgz
- squeak-1.3.0.tgz
- lpad-align-1.1.2.tgz
- meow-3.7.0.tgz
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/apps/commit/6b6da833aa3bd0bc4a04b5931212e0ecf31d7235">6b6da833aa3bd0bc4a04b5931212e0ecf31d7235</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution (minimist): 0.2.1</p>
<p>Direct dependency fix Resolution (mocha): 6.2.3</p><p>Fix Resolution (minimist): 1.2.3</p>
<p>Direct dependency fix Resolution (imagemin-pngquant): 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in minimist tgz minimist tgz autoclosed cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file package json path to vulnerable library node modules mocha node modules minimist package json dependency hierarchy mocha tgz root library mkdirp tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file package json path to vulnerable library node modules meow node modules minimist package json dependency hierarchy imagemin pngquant tgz root library pngquant bin tgz logalot tgz squeak tgz lpad align tgz meow tgz x minimist tgz vulnerable library found in head commit a href found in base branch master vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution minimist direct dependency fix resolution mocha fix resolution minimist direct dependency fix resolution imagemin pngquant step up your open source security game with whitesource
| 0
|
122,233
| 12,146,555,802
|
IssuesEvent
|
2020-04-24 11:23:13
|
assimbly/gateway
|
https://api.github.com/repos/assimbly/gateway
|
closed
|
documentation (user guide) should match gui (navbar)
|
documentation
|
In the navbar are all pages. It's best when the user guides follows the same page from left te right to explain what they do
|
1.0
|
documentation (user guide) should match gui (navbar) - In the navbar are all pages. It's best when the user guides follows the same page from left te right to explain what they do
|
non_test
|
documentation user guide should match gui navbar in the navbar are all pages it s best when the user guides follows the same page from left te right to explain what they do
| 0
|
265,833
| 23,201,588,444
|
IssuesEvent
|
2022-08-01 22:12:22
|
iron-fish/ironfish
|
https://api.github.com/repos/iron-fish/ironfish
|
closed
|
Random characters appear sporadically while updating blocks
|
bug testnet-credit
|
Hi Team,
I ran the following command for installing Iron Fish:
docker run --rm --tty --interactive --network host --volume <home-directory>/.ironfish:/root/.ironfish ghcr.io/iron-fish/ironfish:latest
While updating the blocks, there were patches of random characters appearing in between. May not be a major issue, but I thought I would report it anyway.

Regards,
Pushpraj
|
1.0
|
Random characters appear sporadically while updating blocks - Hi Team,
I ran the following command for installing Iron Fish:
docker run --rm --tty --interactive --network host --volume <home-directory>/.ironfish:/root/.ironfish ghcr.io/iron-fish/ironfish:latest
While updating the blocks, there were patches of random characters appearing in between. May not be a major issue, but I thought I would report it anyway.

Regards,
Pushpraj
|
test
|
random characters appear sporadically while updating blocks hi team i ran the following command for installing iron fish docker run rm tty interactive network host volume ironfish root ironfish ghcr io iron fish ironfish latest while updating the blocks there were patches of random characters appearing in between may not be a major issue but i thought i would report it anyway regards pushpraj
| 1
|
170,089
| 13,172,838,149
|
IssuesEvent
|
2020-08-11 19:10:06
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: django failed
|
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
|
[(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1838926&tab=buildLog) on [release-20.1@6a7ca722a135e21ad04daec3895535969ba5b02c](https://github.com/cockroachdb/cockroach/commits/6a7ca722a135e21ad04daec3895535969ba5b02c):
```
The test failed on branch=release-20.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20200331-1838926/django/run_1
test_runner.go:800: test timed out (10h0m0s)
django.go:192,django.go:212,test_runner.go:753: error with attached stack trace:
main.(*cluster).RunL
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2016
main.registerDjango.func1
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:189
main.registerDjango.func2
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:212
main.(*testRunner).runTest.func2
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753
runtime.goexit
/usr/local/go/src/runtime/asm_amd64.s:1357
- cluster.RunL:
- context canceled
```
<details><summary>More</summary><p>
Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=1838926&tab=artifacts#/django)
Related:
- #46490 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003240059_v20.1.0-beta.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003240059_v20.1.0-beta.4) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46375 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003200044_v20.1.0-beta.3](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003200044_v20.1.0-beta.3) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46321 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003181957_v20.1.0-beta.3](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003181957_v20.1.0-beta.3) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46171 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003161814_v19.2.5](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003161814_v19.2.5) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #45393 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002240854_v20.1.0-beta.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002240854_v20.1.0-beta.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #45299 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-45179](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-45179) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #45063 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-44941](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-44941)
- #44933 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002062122_v19.2.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002062122_v19.2.4)
- #44679 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master)
- #44573 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202001302015_v19.2.3](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202001302015_v19.2.3)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: django failed - [(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1838926&tab=buildLog) on [release-20.1@6a7ca722a135e21ad04daec3895535969ba5b02c](https://github.com/cockroachdb/cockroach/commits/6a7ca722a135e21ad04daec3895535969ba5b02c):
```
The test failed on branch=release-20.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20200331-1838926/django/run_1
test_runner.go:800: test timed out (10h0m0s)
django.go:192,django.go:212,test_runner.go:753: error with attached stack trace:
main.(*cluster).RunL
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2016
main.registerDjango.func1
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:189
main.registerDjango.func2
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:212
main.(*testRunner).runTest.func2
/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753
runtime.goexit
/usr/local/go/src/runtime/asm_amd64.s:1357
- cluster.RunL:
- context canceled
```
<details><summary>More</summary><p>
Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=1838926&tab=artifacts#/django)
Related:
- #46490 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003240059_v20.1.0-beta.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003240059_v20.1.0-beta.4) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46375 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003200044_v20.1.0-beta.3](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003200044_v20.1.0-beta.3) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46321 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003181957_v20.1.0-beta.3](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003181957_v20.1.0-beta.3) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46171 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202003161814_v19.2.5](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202003161814_v19.2.5) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #45393 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002240854_v20.1.0-beta.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002240854_v20.1.0-beta.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #45299 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-45179](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-45179) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #45063 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-44941](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-44941)
- #44933 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002062122_v19.2.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002062122_v19.2.4)
- #44679 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master)
- #44573 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202001302015_v19.2.3](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202001302015_v19.2.3)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
roachtest django failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts django run test runner go test timed out django go django go test runner go error with attached stack trace main cluster runl home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerdjango home agent work go src github com cockroachdb cockroach pkg cmd roachtest django go main registerdjango home agent work go src github com cockroachdb cockroach pkg cmd roachtest django go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s cluster runl context canceled more artifacts related roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed powered by
| 1
|
148,831
| 23,388,615,576
|
IssuesEvent
|
2022-08-11 15:41:35
|
scholokov/long-travel-2
|
https://api.github.com/repos/scholokov/long-travel-2
|
opened
|
Travel: Название блока в 3 строки
|
Design
|
Название блока должно быть в 3 строки вместо двух
Старий Солотвін
Житомирська область, Україна
25.10.2021

|
1.0
|
Travel: Название блока в 3 строки - Название блока должно быть в 3 строки вместо двух
Старий Солотвін
Житомирська область, Україна
25.10.2021

|
non_test
|
travel название блока в строки название блока должно быть в строки вместо двух старий солотвін житомирська область україна
| 0
|
173,631
| 13,434,329,608
|
IssuesEvent
|
2020-09-07 11:11:49
|
SunstriderEmu/BugTracker
|
https://api.github.com/repos/SunstriderEmu/BugTracker
|
closed
|
Daily Dungeon Reset clears progress of normal dungeons and removes ID
|
confirmed core testfix
|
<!--- (**********************************)
(** Fill in the following fields **)
(**********************************)
Issues are for problem only, NOT for asking questions.
--->
**Description:**
When the daily dungeon reset occurs normal dungeons which do not have saved/locked IDs should not remove players current progress.
**Current behaviour:**
At the daily reset timer, normal dungeons are reset removing group progress and sending them to their hearthstone locations.
**Expected behaviour:**
Non-Heroic, Non-Raid dungeons should not be reset during the daily reset process.
**Steps to reproduce the problem:**
1. Be in a dungeon.
2. Wait for the daily reset, this could take up to 24 hours.
3. Be sad you are now in your hearth location instead of the dungeon you are working hard on with your friends.
4. Report to the Github page.
5. Give Nolin some whoppers.
During the above steps I was drinking coffee in my pajamas so you might want to take the same approach to ensure the same result.
**rev. hash/commit:**

|
1.0
|
Daily Dungeon Reset clears progress of normal dungeons and removes ID - <!--- (**********************************)
(** Fill in the following fields **)
(**********************************)
Issues are for problem only, NOT for asking questions.
--->
**Description:**
When the daily dungeon reset occurs normal dungeons which do not have saved/locked IDs should not remove players current progress.
**Current behaviour:**
At the daily reset timer, normal dungeons are reset removing group progress and sending them to their hearthstone locations.
**Expected behaviour:**
Non-Heroic, Non-Raid dungeons should not be reset during the daily reset process.
**Steps to reproduce the problem:**
1. Be in a dungeon.
2. Wait for the daily reset, this could take up to 24 hours.
3. Be sad you are now in your hearth location instead of the dungeon you are working hard on with your friends.
4. Report to the Github page.
5. Give Nolin some whoppers.
During the above steps I was drinking coffee in my pajamas so you might want to take the same approach to ensure the same result.
**rev. hash/commit:**

|
test
|
daily dungeon reset clears progress of normal dungeons and removes id fill in the following fields issues are for problem only not for asking questions description when the daily dungeon reset occurs normal dungeons which do not have saved locked ids should not remove players current progress current behaviour at the daily reset timer normal dungeons are reset removing group progress and sending them to their hearthstone locations expected behaviour non heroic non raid dungeons should not be reset during the daily reset process steps to reproduce the problem be in a dungeon wait for the daily reset this could take up to hours be sad you are now in your hearth location instead of the dungeon you are working hard on with your friends report to the github page give nolin some whoppers during the above steps i was drinking coffee in my pajamas so you might want to take the same approach to ensure the same result rev hash commit
| 1
|
111,510
| 9,533,401,847
|
IssuesEvent
|
2019-04-29 21:11:19
|
chamilo/chamilo-lms
|
https://api.github.com/repos/chamilo/chamilo-lms
|
closed
|
Problem creating a certificate.
|
Bug Requires testing/validation
|
Hello.
When creating a course in Learning paths.
You click certificate.
After creating it is impossible to edit.
tested at https://11.chamilo.org/
Greetings.


|
1.0
|
Problem creating a certificate. - Hello.
When creating a course in Learning paths.
You click certificate.
After creating it is impossible to edit.
tested at https://11.chamilo.org/
Greetings.


|
test
|
problem creating a certificate hello when creating a course in learning paths you click certificate after creating it is impossible to edit tested at greetings
| 1
|
20,421
| 4,543,551,623
|
IssuesEvent
|
2016-09-10 06:16:59
|
BVLC/caffe
|
https://api.github.com/repos/BVLC/caffe
|
closed
|
Net Surgeries
|
documentation
|
The [editing model parameters](http://nbviewer.ipython.org/github/BVLC/caffe/blob/master/examples/net_surgery.ipynb) example only hints at the possibilities of net surgery. To make it more useful and better explain pycaffe along the way it should have:
- transplanting custom filters (Gaussian, Sobel, bilinear, ...) and classifier weights (like a separately learned SVM)
- combining layers from different nets / making an ensemble in a single model
- grafting new channels to a model to augment the number of classes
- ~~switch the channel order of conv1, or convert between grayscale and color~~
These are all possible as-is, but it takes know-how.
|
1.0
|
Net Surgeries - The [editing model parameters](http://nbviewer.ipython.org/github/BVLC/caffe/blob/master/examples/net_surgery.ipynb) example only hints at the possibilities of net surgery. To make it more useful and better explain pycaffe along the way it should have:
- transplanting custom filters (Gaussian, Sobel, bilinear, ...) and classifier weights (like a separately learned SVM)
- combining layers from different nets / making an ensemble in a single model
- grafting new channels to a model to augment the number of classes
- ~~switch the channel order of conv1, or convert between grayscale and color~~
These are all possible as-is, but it takes know-how.
|
non_test
|
net surgeries the example only hints at the possibilities of net surgery to make it more useful and better explain pycaffe along the way it should have transplanting custom filters gaussian sobel bilinear and classifier weights like a separately learned svm combining layers from different nets making an ensemble in a single model grafting new channels to a model to augment the number of classes switch the channel order of or convert between grayscale and color these are all possible as is but it takes know how
| 0
|
60,702
| 6,713,437,011
|
IssuesEvent
|
2017-10-13 13:28:44
|
MajkiIT/polish-ads-filter
|
https://api.github.com/repos/MajkiIT/polish-ads-filter
|
reopened
|
cda-hd.pl
|
dodać regex reguły gotowe/testowanie
|
@MajkiIT looknie sobie na byle jaki film z flashx. nie otwiera sie tobie nic extra poza regexem?
http://cda-hd.pl/20944/dom-wygranych-the-house-2017-online/
|
1.0
|
cda-hd.pl - @MajkiIT looknie sobie na byle jaki film z flashx. nie otwiera sie tobie nic extra poza regexem?
http://cda-hd.pl/20944/dom-wygranych-the-house-2017-online/
|
test
|
cda hd pl majkiit looknie sobie na byle jaki film z flashx nie otwiera sie tobie nic extra poza regexem
| 1
|
204,415
| 15,441,123,027
|
IssuesEvent
|
2021-03-08 05:13:23
|
trevorNgo/Measure2.0
|
https://api.github.com/repos/trevorNgo/Measure2.0
|
opened
|
CS4ZP6 Tester Feedback: Archive form is non-functional
|
tester
|
Environment
- Google Chrome
- Windows 10
Describe the feedback/issue.
Completing the `Archive Year Term` form does not do anything:

Priority
HIGH
Please list the reproduction steps
- Login as `Admin`
- Navigate to and complete the `Archive Year Term` form
- Check results
|
1.0
|
CS4ZP6 Tester Feedback: Archive form is non-functional - Environment
- Google Chrome
- Windows 10
Describe the feedback/issue.
Completing the `Archive Year Term` form does not do anything:

Priority
HIGH
Please list the reproduction steps
- Login as `Admin`
- Navigate to and complete the `Archive Year Term` form
- Check results
|
test
|
tester feedback archive form is non functional environment google chrome windows describe the feedback issue completing the archive year term form does not do anything priority high please list the reproduction steps login as admin navigate to and complete the archive year term form check results
| 1
|
7,956
| 4,111,699,594
|
IssuesEvent
|
2016-06-07 07:34:33
|
youjustgo/ng2-bingmaps
|
https://api.github.com/repos/youjustgo/ng2-bingmaps
|
closed
|
Set up demo inside the repository
|
build
|
We would like to have a demo folder inside the repository, for easier testing. This should reference the file generated by the build (dist folder)
|
1.0
|
Set up demo inside the repository - We would like to have a demo folder inside the repository, for easier testing. This should reference the file generated by the build (dist folder)
|
non_test
|
set up demo inside the repository we would like to have a demo folder inside the repository for easier testing this should reference the file generated by the build dist folder
| 0
|
254,345
| 21,781,368,677
|
IssuesEvent
|
2022-05-13 19:20:57
|
ruffle-rs/ruffle
|
https://api.github.com/repos/ruffle-rs/ruffle
|
opened
|
Add more image tests
|
good first issue tests
|
@Aaron1011 was kind enough to add the ability for rendering test images, we should add some more of these, some of which may have caught recent regressions.
* Fill styles (gradients, bitmaps, w/ transforms)
* Line styles (including all linestyle2 options, caps, joins, fills)
* Color transforms (including <0% and >100% testing premultiplied alpha)
* Masking (including nested masks, AS masks)
* Text (static, edit texts, outline and device fonts)
* Drawing API
* BitmapData (and associated methods)
* Blend modes (when implemented)
* Filters (when implemented)
|
1.0
|
Add more image tests - @Aaron1011 was kind enough to add the ability for rendering test images, we should add some more of these, some of which may have caught recent regressions.
* Fill styles (gradients, bitmaps, w/ transforms)
* Line styles (including all linestyle2 options, caps, joins, fills)
* Color transforms (including <0% and >100% testing premultiplied alpha)
* Masking (including nested masks, AS masks)
* Text (static, edit texts, outline and device fonts)
* Drawing API
* BitmapData (and associated methods)
* Blend modes (when implemented)
* Filters (when implemented)
|
test
|
add more image tests was kind enough to add the ability for rendering test images we should add some more of these some of which may have caught recent regressions fill styles gradients bitmaps w transforms line styles including all options caps joins fills color transforms including testing premultiplied alpha masking including nested masks as masks text static edit texts outline and device fonts drawing api bitmapdata and associated methods blend modes when implemented filters when implemented
| 1
|
292,332
| 25,206,566,101
|
IssuesEvent
|
2022-11-13 19:00:16
|
MinhazMurks/Bannerlord.Tweaks
|
https://api.github.com/repos/MinhazMurks/Bannerlord.Tweaks
|
opened
|
Test Town Militia Barracks Production Level 3
|
testing
|
Test to see if tweak: "Town Militia Barracks Production Level 3" works
|
1.0
|
Test Town Militia Barracks Production Level 3 - Test to see if tweak: "Town Militia Barracks Production Level 3" works
|
test
|
test town militia barracks production level test to see if tweak town militia barracks production level works
| 1
|
209,600
| 16,044,127,796
|
IssuesEvent
|
2021-04-22 11:40:32
|
Kiryakor/QA
|
https://api.github.com/repos/Kiryakor/QA
|
opened
|
QA-3: необходимо протестировать Апи
|
test
|
# Необходимо протестировать Апи для работы с бд и написать тесты
### [ссылка на dev task](https://github.com/Kiryakor/QA/issues/5)
|
1.0
|
QA-3: необходимо протестировать Апи - # Необходимо протестировать Апи для работы с бд и написать тесты
### [ссылка на dev task](https://github.com/Kiryakor/QA/issues/5)
|
test
|
qa необходимо протестировать апи необходимо протестировать апи для работы с бд и написать тесты
| 1
|
51,237
| 6,152,002,628
|
IssuesEvent
|
2017-06-28 05:28:56
|
druid-io/druid
|
https://api.github.com/repos/druid-io/druid
|
closed
|
Transient failures of integration tests
|
Testing
|
Please edit messages with failure "snapshots" by adding links to failing builds, to understand which tests fail most often.
|
1.0
|
Transient failures of integration tests - Please edit messages with failure "snapshots" by adding links to failing builds, to understand which tests fail most often.
|
test
|
transient failures of integration tests please edit messages with failure snapshots by adding links to failing builds to understand which tests fail most often
| 1
|
86,791
| 10,518,398,666
|
IssuesEvent
|
2019-09-29 10:29:02
|
Jugendhackt/parteiduell-frontend
|
https://api.github.com/repos/Jugendhackt/parteiduell-frontend
|
closed
|
README
|
documentation good first issue help wanted
|
Erstellt doch bitte noch ein schönes README, in welchem erklärt wird, wie man das Frontend lokal aufsetzten kann und wie man am Projekt teilhaben kann.
|
1.0
|
README - Erstellt doch bitte noch ein schönes README, in welchem erklärt wird, wie man das Frontend lokal aufsetzten kann und wie man am Projekt teilhaben kann.
|
non_test
|
readme erstellt doch bitte noch ein schönes readme in welchem erklärt wird wie man das frontend lokal aufsetzten kann und wie man am projekt teilhaben kann
| 0
|
122,661
| 10,229,003,634
|
IssuesEvent
|
2019-08-17 08:31:44
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Rephrase the error text for file not found error
|
Area/StandardLibs BetaTesting Type/Improvement
|
Consider the following:
```ballerina
system:FileInfo[] fileList = checkpanic system:readDir("invalid/path");
```
The error given for the above is:
```
error: error {ballerina/system}InvalidOperationError message=File doesn't exist in path ~/Downloads cause=null
```
Can't we just simply say `File not found: <file_path>`? "File doesn't exist in path" sounds odd.
|
1.0
|
Rephrase the error text for file not found error - Consider the following:
```ballerina
system:FileInfo[] fileList = checkpanic system:readDir("invalid/path");
```
The error given for the above is:
```
error: error {ballerina/system}InvalidOperationError message=File doesn't exist in path ~/Downloads cause=null
```
Can't we just simply say `File not found: <file_path>`? "File doesn't exist in path" sounds odd.
|
test
|
rephrase the error text for file not found error consider the following ballerina system fileinfo filelist checkpanic system readdir invalid path the error given for the above is error error ballerina system invalidoperationerror message file doesn t exist in path downloads cause null can t we just simply say file not found file doesn t exist in path sounds odd
| 1
|
181,769
| 14,074,544,803
|
IssuesEvent
|
2020-11-04 07:31:39
|
OpenMined/PySyft
|
https://api.github.com/repos/OpenMined/PySyft
|
closed
|
Add torch.Tensor.permute to allowlist and test suite
|
Priority: 2 - High :cold_sweat: Severity: 3 - Medium :unamused: Status: Available :wave: Type: New Feature :heavy_plus_sign: Type: Testing :test_tube:
|
# Description
This issue is a part of Syft 0.3.0 Epic 2: https://github.com/OpenMined/PySyft/issues/3696
In this issue, you will be adding support for remote execution of the torch.Tensor.permute
method or property. This might be a really small project (literally a one-liner) or
it might require adding significant functionality to PySyft OR to the testing suite
in order to make sure the feature is both functional and tested.
## Step 0: Run tests and ./scripts/pre_commit.sh
Before you get started with this project, let's make sure you have everything building and testing
correctly. Clone the codebase and run:
```pip uninstall syft```
followed by
```pip install -e .```
Then run the pre-commit file (which will also run the tests)
```./scripts/pre_commit.sh```
If all of these tests pass, continue on. If not, make sure you have all the
dependencies in requirements.txt installed, etc.
## Step 1: Uncomment your method in the allowlist.py file
Inside [allowlist.py](https://github.com/OpenMined/PySyft/blob/syft_0.3.0/src/syft/lib/torch/allowlist.py) you will find a huge dictionary of methods. Find your method and uncomment the line its on. At the time
of writing this Issue (WARNING: THIS MAY HAVE CHANGED) the dictionary maps from the
string name of the method (in your case 'torch.Tensor.permute') to the string representation
of the type the method returns.
## Step 2: Run Unit Tests
Run the following:
```python setup.py test```
And wait to see if some of the tests fail. Why might the tests fail now? I'm so glad you asked!
https://github.com/OpenMined/PySyft/blob/syft_0.3.0/tests/syft/lib/torch/tensor/tensor_remote_method_api_suite_test.py
In this file you'll find the torch method test suite. It AUTOMATICALLY loads all methods
from the allowlist.py file you modified in the previous step. It attempts to test them.
# Step 3: If you get a Failing Test
If you get a failing test, this could be for one of a few reasons:
### Reason 1 - The testing suite passed in non-compatible arguments
The testing suite is pretty dumb. It literally just has a permutation of possible
arguments to pass into every method on torch tensors. So, if one of those permutations
doesn't work for your method (aka... perhaps it tries to call your method without
any arguments but torch.Tensor.permute actually requires some) then the test will
fail if the error hasn't been seen before.
If this happens - don't worry! Just look inside the only test in that file and look
for the huge lists of error types to ignore. Add your error to the list and keep
going!!!
*WARNING:* make sure that the testing suite actually tests your method via remote
execution once you've gotten all the tests passing. Aka - if the testing suite
doesn't have ANY matching argument permutations for your method, then you're
literally creating a bunch of unit tests that do absolutely nothing. If this is the
case, then ADD MORE ARGUMENT TYPES TO THE TESTING SUITE so that your argument
gets run via remote execution. DO NOT CLOSE THIS ISSUE until you can verify that
torch.Tensor.permute is actually executed remotely inside of a unit tests (and not
skipped). Aka - at least one of the test_all_allowlisted_tensor_methods_work_remotely_on_all_types
unit tests with your method should run ALL THE WAY TO THE END (instead of skipping
the last part.)
*Note:* adding another argument type might require some serialization work if
we don't support arguments of that type yet. If so, this is your job to add it
to the protobuf files in order to close this issue!
### Reason 2 - torch.Tensor.permute returns a non-supported type
If this happens, you've got a little bit of work in front of you. We don't have
pointer objects to very many remote object types. So, if your method returns anything
other than a single tensor, you probably need to add support for the type it returns
(Such as a bool, None, int, or other types).
*IMPORTANT:* do NOT return the value itself to the end user!!! Return a pointer object
to that type!
*NOTE:* at the time of writing - there are several core pieces of Syft not yet working
to allow you to return any type other than a torch tensor. If you're not comfortable
investigating what those might be - skip this issue and try again later once
someone else has solved these issues.
### Reason 3 - There's something else broken
Chase those stack traces! Talk to friends in Slack. Look at how other methods are supported.
This is a challenging project in a fast moving codebase!
And don't forget - if this project seems to complex - there are plenty of others that
might be easier.
|
2.0
|
Add torch.Tensor.permute to allowlist and test suite -
# Description
This issue is a part of Syft 0.3.0 Epic 2: https://github.com/OpenMined/PySyft/issues/3696
In this issue, you will be adding support for remote execution of the torch.Tensor.permute
method or property. This might be a really small project (literally a one-liner) or
it might require adding significant functionality to PySyft OR to the testing suite
in order to make sure the feature is both functional and tested.
## Step 0: Run tests and ./scripts/pre_commit.sh
Before you get started with this project, let's make sure you have everything building and testing
correctly. Clone the codebase and run:
```pip uninstall syft```
followed by
```pip install -e .```
Then run the pre-commit file (which will also run the tests)
```./scripts/pre_commit.sh```
If all of these tests pass, continue on. If not, make sure you have all the
dependencies in requirements.txt installed, etc.
## Step 1: Uncomment your method in the allowlist.py file
Inside [allowlist.py](https://github.com/OpenMined/PySyft/blob/syft_0.3.0/src/syft/lib/torch/allowlist.py) you will find a huge dictionary of methods. Find your method and uncomment the line its on. At the time
of writing this Issue (WARNING: THIS MAY HAVE CHANGED) the dictionary maps from the
string name of the method (in your case 'torch.Tensor.permute') to the string representation
of the type the method returns.
## Step 2: Run Unit Tests
Run the following:
```python setup.py test```
And wait to see if some of the tests fail. Why might the tests fail now? I'm so glad you asked!
https://github.com/OpenMined/PySyft/blob/syft_0.3.0/tests/syft/lib/torch/tensor/tensor_remote_method_api_suite_test.py
In this file you'll find the torch method test suite. It AUTOMATICALLY loads all methods
from the allowlist.py file you modified in the previous step. It attempts to test them.
# Step 3: If you get a Failing Test
If you get a failing test, this could be for one of a few reasons:
### Reason 1 - The testing suite passed in non-compatible arguments
The testing suite is pretty dumb. It literally just has a permutation of possible
arguments to pass into every method on torch tensors. So, if one of those permutations
doesn't work for your method (aka... perhaps it tries to call your method without
any arguments but torch.Tensor.permute actually requires some) then the test will
fail if the error hasn't been seen before.
If this happens - don't worry! Just look inside the only test in that file and look
for the huge lists of error types to ignore. Add your error to the list and keep
going!!!
*WARNING:* make sure that the testing suite actually tests your method via remote
execution once you've gotten all the tests passing. Aka - if the testing suite
doesn't have ANY matching argument permutations for your method, then you're
literally creating a bunch of unit tests that do absolutely nothing. If this is the
case, then ADD MORE ARGUMENT TYPES TO THE TESTING SUITE so that your argument
gets run via remote execution. DO NOT CLOSE THIS ISSUE until you can verify that
torch.Tensor.permute is actually executed remotely inside of a unit tests (and not
skipped). Aka - at least one of the test_all_allowlisted_tensor_methods_work_remotely_on_all_types
unit tests with your method should run ALL THE WAY TO THE END (instead of skipping
the last part.)
*Note:* adding another argument type might require some serialization work if
we don't support arguments of that type yet. If so, this is your job to add it
to the protobuf files in order to close this issue!
### Reason 2 - torch.Tensor.permute returns a non-supported type
If this happens, you've got a little bit of work in front of you. We don't have
pointer objects to very many remote object types. So, if your method returns anything
other than a single tensor, you probably need to add support for the type it returns
(Such as a bool, None, int, or other types).
*IMPORTANT:* do NOT return the value itself to the end user!!! Return a pointer object
to that type!
*NOTE:* at the time of writing - there are several core pieces of Syft not yet working
to allow you to return any type other than a torch tensor. If you're not comfortable
investigating what those might be - skip this issue and try again later once
someone else has solved these issues.
### Reason 3 - There's something else broken
Chase those stack traces! Talk to friends in Slack. Look at how other methods are supported.
This is a challenging project in a fast moving codebase!
And don't forget - if this project seems to complex - there are plenty of others that
might be easier.
|
test
|
add torch tensor permute to allowlist and test suite description this issue is a part of syft epic in this issue you will be adding support for remote execution of the torch tensor permute method or property this might be a really small project literally a one liner or it might require adding significant functionality to pysyft or to the testing suite in order to make sure the feature is both functional and tested step run tests and scripts pre commit sh before you get started with this project let s make sure you have everything building and testing correctly clone the codebase and run pip uninstall syft followed by pip install e then run the pre commit file which will also run the tests scripts pre commit sh if all of these tests pass continue on if not make sure you have all the dependencies in requirements txt installed etc step uncomment your method in the allowlist py file inside you will find a huge dictionary of methods find your method and uncomment the line its on at the time of writing this issue warning this may have changed the dictionary maps from the string name of the method in your case torch tensor permute to the string representation of the type the method returns step run unit tests run the following python setup py test and wait to see if some of the tests fail why might the tests fail now i m so glad you asked in this file you ll find the torch method test suite it automatically loads all methods from the allowlist py file you modified in the previous step it attempts to test them step if you get a failing test if you get a failing test this could be for one of a few reasons reason the testing suite passed in non compatible arguments the testing suite is pretty dumb it literally just has a permutation of possible arguments to pass into every method on torch tensors so if one of those permutations doesn t work for your method aka perhaps it tries to call your method without any arguments but torch tensor permute actually requires some then the test will fail if the error hasn t been seen before if this happens don t worry just look inside the only test in that file and look for the huge lists of error types to ignore add your error to the list and keep going warning make sure that the testing suite actually tests your method via remote execution once you ve gotten all the tests passing aka if the testing suite doesn t have any matching argument permutations for your method then you re literally creating a bunch of unit tests that do absolutely nothing if this is the case then add more argument types to the testing suite so that your argument gets run via remote execution do not close this issue until you can verify that torch tensor permute is actually executed remotely inside of a unit tests and not skipped aka at least one of the test all allowlisted tensor methods work remotely on all types unit tests with your method should run all the way to the end instead of skipping the last part note adding another argument type might require some serialization work if we don t support arguments of that type yet if so this is your job to add it to the protobuf files in order to close this issue reason torch tensor permute returns a non supported type if this happens you ve got a little bit of work in front of you we don t have pointer objects to very many remote object types so if your method returns anything other than a single tensor you probably need to add support for the type it returns such as a bool none int or other types important do not return the value itself to the end user return a pointer object to that type note at the time of writing there are several core pieces of syft not yet working to allow you to return any type other than a torch tensor if you re not comfortable investigating what those might be skip this issue and try again later once someone else has solved these issues reason there s something else broken chase those stack traces talk to friends in slack look at how other methods are supported this is a challenging project in a fast moving codebase and don t forget if this project seems to complex there are plenty of others that might be easier
| 1
|
281,525
| 30,888,881,142
|
IssuesEvent
|
2023-08-04 01:57:39
|
hshivhare67/kernel_v4.1.15_CVE-2019-10220
|
https://api.github.com/repos/hshivhare67/kernel_v4.1.15_CVE-2019-10220
|
reopened
|
CVE-2020-27170 (Medium) detected in linuxlinux-4.4.302
|
Mend: dependency security vulnerability
|
## CVE-2020-27170 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.1.15_CVE-2019-10220/commit/85c655ae4a38840b0edb9284bb64ec65b27331dc">85c655ae4a38840b0edb9284bb64ec65b27331dc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.11.8. kernel/bpf/verifier.c performs undesirable out-of-bounds speculation on pointer arithmetic, leading to side-channel attacks that defeat Spectre mitigations and obtain sensitive information from kernel memory, aka CID-f232326f6966. This affects pointer types that do not define a ptr_limit.
<p>Publish Date: 2021-03-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-27170>CVE-2020-27170</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-27170">https://nvd.nist.gov/vuln/detail/CVE-2020-27170</a></p>
<p>Release Date: 2021-03-20</p>
<p>Fix Resolution: kernel-rt-debug-devel - 3.10.0-1160.31.1.rt56.1169;python-perf - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-debug - 3.10.0-1160.31.1.rt56.1169;kernel-rt-devel - 3.10.0-1160.31.1.rt56.1169;kernel-debuginfo-common-ppc64 - 3.10.0-1160.31.1;kernel-tools - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-trace-kvm - 3.10.0-1160.31.1.rt56.1169;kernel-tools-libs-devel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-tools-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-doc - 3.10.0-1160.31.1.rt56.1169;kernel-rt-kvm - 3.10.0-1160.31.1.rt56.1169;python-perf-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-headers - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-trace - 3.10.0-1160.31.1.rt56.1169;kernel-debuginfo-common-x86_64 - 3.10.0-1160.31.1;kernel-rt - 3.10.0-1160.31.1.rt56.1169,3.10.0-1160.31.1.rt56.1169;kernel-doc - 3.10.0-1160.31.1;kernel-abi-whitelists - 3.10.0-1160.31.1;kernel-rt-trace-devel - 3.10.0-1160.31.1.rt56.1169;kernel-rt-debug-kvm - 3.10.0-1160.31.1.rt56.1169;kernel-bootwrapper - 3.10.0-1160.31.1,3.10.0-1160.31.1;perf - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debug-devel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;bpftool - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-tools-libs - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;perf-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debug-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debug - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-devel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;bpftool-debuginfo - 3.10.0-1160.31.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-27170 (Medium) detected in linuxlinux-4.4.302 - ## CVE-2020-27170 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.1.15_CVE-2019-10220/commit/85c655ae4a38840b0edb9284bb64ec65b27331dc">85c655ae4a38840b0edb9284bb64ec65b27331dc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.11.8. kernel/bpf/verifier.c performs undesirable out-of-bounds speculation on pointer arithmetic, leading to side-channel attacks that defeat Spectre mitigations and obtain sensitive information from kernel memory, aka CID-f232326f6966. This affects pointer types that do not define a ptr_limit.
<p>Publish Date: 2021-03-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-27170>CVE-2020-27170</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-27170">https://nvd.nist.gov/vuln/detail/CVE-2020-27170</a></p>
<p>Release Date: 2021-03-20</p>
<p>Fix Resolution: kernel-rt-debug-devel - 3.10.0-1160.31.1.rt56.1169;python-perf - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-debug - 3.10.0-1160.31.1.rt56.1169;kernel-rt-devel - 3.10.0-1160.31.1.rt56.1169;kernel-debuginfo-common-ppc64 - 3.10.0-1160.31.1;kernel-tools - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-trace-kvm - 3.10.0-1160.31.1.rt56.1169;kernel-tools-libs-devel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-tools-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-doc - 3.10.0-1160.31.1.rt56.1169;kernel-rt-kvm - 3.10.0-1160.31.1.rt56.1169;python-perf-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-headers - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-rt-trace - 3.10.0-1160.31.1.rt56.1169;kernel-debuginfo-common-x86_64 - 3.10.0-1160.31.1;kernel-rt - 3.10.0-1160.31.1.rt56.1169,3.10.0-1160.31.1.rt56.1169;kernel-doc - 3.10.0-1160.31.1;kernel-abi-whitelists - 3.10.0-1160.31.1;kernel-rt-trace-devel - 3.10.0-1160.31.1.rt56.1169;kernel-rt-debug-kvm - 3.10.0-1160.31.1.rt56.1169;kernel-bootwrapper - 3.10.0-1160.31.1,3.10.0-1160.31.1;perf - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debug-devel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;bpftool - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-tools-libs - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;perf-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debug-debuginfo - 3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-debug - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel-devel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;kernel - 3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1,3.10.0-1160.31.1;bpftool-debuginfo - 3.10.0-1160.31.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files kernel bpf verifier c kernel bpf verifier c vulnerability details an issue was discovered in the linux kernel before kernel bpf verifier c performs undesirable out of bounds speculation on pointer arithmetic leading to side channel attacks that defeat spectre mitigations and obtain sensitive information from kernel memory aka cid this affects pointer types that do not define a ptr limit publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution kernel rt debug devel python perf kernel rt debug kernel rt devel kernel debuginfo common kernel tools kernel rt trace kvm kernel tools libs devel kernel tools debuginfo kernel rt doc kernel rt kvm python perf debuginfo kernel headers kernel rt trace kernel debuginfo common kernel rt kernel doc kernel abi whitelists kernel rt trace devel kernel rt debug kvm kernel bootwrapper perf kernel debuginfo kernel debug devel bpftool kernel tools libs perf debuginfo kernel debug debuginfo kernel debug kernel devel kernel bpftool debuginfo step up your open source security game with mend
| 0
|
731,440
| 25,216,426,397
|
IssuesEvent
|
2022-11-14 09:30:07
|
OpenSpace/OpenSpace
|
https://api.github.com/repos/OpenSpace/OpenSpace
|
closed
|
Flying to "1 AU Size Comparison Circle" freezes application
|
Type: Bug Priority: Critical Component: Interaction Feature: Camera Paths
|
To reproduce:
1. Add exoplanet system "11 Umi"
2. Find "1 AU Size Comparison Circle" in Navigation list and click the "fly to" button
One beautiful day, the camera path system is going to be free from bugs...
I believe the actual problem might be flying using the avoid collision curve to objects without bounding spheres
(Note: the comparison circles should get a bounding sphere computed!)
|
1.0
|
Flying to "1 AU Size Comparison Circle" freezes application - To reproduce:
1. Add exoplanet system "11 Umi"
2. Find "1 AU Size Comparison Circle" in Navigation list and click the "fly to" button
One beautiful day, the camera path system is going to be free from bugs...
I believe the actual problem might be flying using the avoid collision curve to objects without bounding spheres
(Note: the comparison circles should get a bounding sphere computed!)
|
non_test
|
flying to au size comparison circle freezes application to reproduce add exoplanet system umi find au size comparison circle in navigation list and click the fly to button one beautiful day the camera path system is going to be free from bugs i believe the actual problem might be flying using the avoid collision curve to objects without bounding spheres note the comparison circles should get a bounding sphere computed
| 0
|
134,060
| 10,879,721,221
|
IssuesEvent
|
2019-11-17 04:31:00
|
variar/klogg
|
https://api.github.com/repos/variar/klogg
|
closed
|
Crashes on startup when prior file no longer exists
|
bug case for CI ready for testing
|
Starting with a release after v501, (I verified problem exists with 504, 505, 506), Klogg crashes when reopening files from last session and one of the files no longer exists.
Prior, Klogg successfully opened and the tab existed with an empty content pane.
This was useful because of "clean builds" removing log files and Klogg keep them in the tab list (and startup successfully!).
|
1.0
|
Crashes on startup when prior file no longer exists - Starting with a release after v501, (I verified problem exists with 504, 505, 506), Klogg crashes when reopening files from last session and one of the files no longer exists.
Prior, Klogg successfully opened and the tab existed with an empty content pane.
This was useful because of "clean builds" removing log files and Klogg keep them in the tab list (and startup successfully!).
|
test
|
crashes on startup when prior file no longer exists starting with a release after i verified problem exists with klogg crashes when reopening files from last session and one of the files no longer exists prior klogg successfully opened and the tab existed with an empty content pane this was useful because of clean builds removing log files and klogg keep them in the tab list and startup successfully
| 1
|
65,830
| 27,244,439,351
|
IssuesEvent
|
2023-02-22 00:03:30
|
microsoft/vscode-cpptools
|
https://api.github.com/repos/microsoft/vscode-cpptools
|
opened
|
#include <test.h> completion to test2.h creates `test2.h""`
|
bug Language Service Feature: Auto-complete
|
With test.h and test2.h and
`#include "test.h"`
invoking include completion to replace it with test2.h creates `test2.h""`
|
1.0
|
#include <test.h> completion to test2.h creates `test2.h""` - With test.h and test2.h and
`#include "test.h"`
invoking include completion to replace it with test2.h creates `test2.h""`
|
non_test
|
include completion to h creates h with test h and h and include test h invoking include completion to replace it with h creates h
| 0
|
277,062
| 24,045,973,107
|
IssuesEvent
|
2022-09-16 08:22:13
|
Azure/ResourceModules
|
https://api.github.com/repos/Azure/ResourceModules
|
opened
|
[Feature Request]: Updated dependencies approach - Test formatting
|
enhancement [cat] testing blocked [cat] needs further discussion
|
### Description
This issue is about agreeing on the format to apply to module test templates (deploy.test.bicep) and implement the changes as a bulk edit once all related PRs will be merged.
In particular, section names and spacing e.g. use `// Dependencies //` instead of generic `// Deployments //` for the dependencies section.
The issue is blocked by #1791 or at least by the merge of all currently open PRs.
|
1.0
|
[Feature Request]: Updated dependencies approach - Test formatting - ### Description
This issue is about agreeing on the format to apply to module test templates (deploy.test.bicep) and implement the changes as a bulk edit once all related PRs will be merged.
In particular, section names and spacing e.g. use `// Dependencies //` instead of generic `// Deployments //` for the dependencies section.
The issue is blocked by #1791 or at least by the merge of all currently open PRs.
|
test
|
updated dependencies approach test formatting description this issue is about agreeing on the format to apply to module test templates deploy test bicep and implement the changes as a bulk edit once all related prs will be merged in particular section names and spacing e g use dependencies instead of generic deployments for the dependencies section the issue is blocked by or at least by the merge of all currently open prs
| 1
|
383,474
| 26,551,273,195
|
IssuesEvent
|
2023-01-20 08:00:57
|
supabase/supabase
|
https://api.github.com/repos/supabase/supabase
|
closed
|
Returning data sets requires further instruction
|
documentation
|
# Returning data sets is not debuggable?
## Link
https://supabase.com/docs/guides/database/functions#returning-data-sets
## Describe the problem
The current example leads to the following problem when I try to use it on my own dbs:
"return type mismatch in function declared to return posts"
The problem is that this makes sense. In the example it looks like the db is defined as a type. But this isn't done in the example, so there's no way to follow this. I can't even find a way to test what the datatype of my return value is when it's in a set?
## Describe the improvement
Either define what datatype is used and how. Or how to debug this error.
This is my example code that creates the error.
```
CREATE OR REPLACE FUNCTION list_unique_domains()
returns setof posts
language sql
as $$
SELECT domain
FROM posts
GROUP BY domain;
$$;
```
|
1.0
|
Returning data sets requires further instruction - # Returning data sets is not debuggable?
## Link
https://supabase.com/docs/guides/database/functions#returning-data-sets
## Describe the problem
The current example leads to the following problem when I try to use it on my own dbs:
"return type mismatch in function declared to return posts"
The problem is that this makes sense. In the example it looks like the db is defined as a type. But this isn't done in the example, so there's no way to follow this. I can't even find a way to test what the datatype of my return value is when it's in a set?
## Describe the improvement
Either define what datatype is used and how. Or how to debug this error.
This is my example code that creates the error.
```
CREATE OR REPLACE FUNCTION list_unique_domains()
returns setof posts
language sql
as $$
SELECT domain
FROM posts
GROUP BY domain;
$$;
```
|
non_test
|
returning data sets requires further instruction returning data sets is not debuggable link describe the problem the current example leads to the following problem when i try to use it on my own dbs return type mismatch in function declared to return posts the problem is that this makes sense in the example it looks like the db is defined as a type but this isn t done in the example so there s no way to follow this i can t even find a way to test what the datatype of my return value is when it s in a set describe the improvement either define what datatype is used and how or how to debug this error this is my example code that creates the error create or replace function list unique domains returns setof posts language sql as select domain from posts group by domain
| 0
|
430,474
| 12,453,528,831
|
IssuesEvent
|
2020-05-27 13:59:01
|
ISDM-G4/G4-ISDM-Project
|
https://api.github.com/repos/ISDM-G4/G4-ISDM-Project
|
opened
|
US#112 - As a relationship manager I want outbound calls to be automatically dialed by the system so that I can focus on sales and not dailing numbers from a list.
|
high-priority
|
As a relationship manager I want outbound calls to be automatically dialed by the system so that I can focus on sales and not dailing numbers from a list.
|
1.0
|
US#112 - As a relationship manager I want outbound calls to be automatically dialed by the system so that I can focus on sales and not dailing numbers from a list. - As a relationship manager I want outbound calls to be automatically dialed by the system so that I can focus on sales and not dailing numbers from a list.
|
non_test
|
us as a relationship manager i want outbound calls to be automatically dialed by the system so that i can focus on sales and not dailing numbers from a list as a relationship manager i want outbound calls to be automatically dialed by the system so that i can focus on sales and not dailing numbers from a list
| 0
|
26,840
| 4,249,572,514
|
IssuesEvent
|
2016-07-08 00:45:30
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
opened
|
kubernetes-e2e-gke: broken test run
|
kind/flake priority/P2 team/test-infra
|
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke/10735/
Multiple broken tests:
Failed: [k8s.io] Kubectl client [k8s.io] Update Demo should do a rolling update of a replication controller [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:233
Expected error:
<*errors.errorString | 0xc8207c7f70>: {
s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-f1wr3] [] 0xc820ae8ae0 Created update-demo-kitten\n Error from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)\n [] <nil> 0xc820ae9100 exit status 1 <nil> true [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba0a8 0xc8200ba130 0xc8200ba548] [0xa9ec00 0xa9ed60 0xa9ed60] 0xc820d6a6c0}:\nCommand stdout:\nCreated update-demo-kitten\n\nstderr:\nError from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)\n\nerror:\nexit status 1\n",
}
Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-f1wr3] [] 0xc820ae8ae0 Created update-demo-kitten
Error from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)
[] <nil> 0xc820ae9100 exit status 1 <nil> true [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba0a8 0xc8200ba130 0xc8200ba548] [0xa9ec00 0xa9ed60 0xa9ed60] 0xc820d6a6c0}:
Command stdout:
Created update-demo-kitten
stderr:
Error from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)
error:
exit status 1
not to have occurred
```
Issues about this test specifically: #26425 #26715
Failed: [k8s.io] Kubectl client [k8s.io] Simple pod should support exec through an HTTP proxy {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:279
Expected error:
<*errors.errorString | 0xc8202aa550>: {
s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-4q19c] [] <nil> the server does not allow access to the requested resource (get services)\n [] <nil> 0xc82072a7c0 exit status 1 <nil> true [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081d0 0xc820108300] [0xa9ed60 0xa9ed60] 0xc820d062a0}:\nCommand stdout:\n\nstderr:\nthe server does not allow access to the requested resource (get services)\n\nerror:\nexit status 1\n",
}
Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-4q19c] [] <nil> the server does not allow access to the requested resource (get services)
[] <nil> 0xc82072a7c0 exit status 1 <nil> true [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081d0 0xc820108300] [0xa9ed60 0xa9ed60] 0xc820d062a0}:
Command stdout:
stderr:
the server does not allow access to the requested resource (get services)
error:
exit status 1
not to have occurred
```
Issues about this test specifically: #27156
Failed: [k8s.io] Services should serve a basic endpoint from pods [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Jul 7 17:24:27.653: Couldn't delete ns "e2e-tests-services-w86n1": the server does not allow access to the requested resource (delete namespaces e2e-tests-services-w86n1)
```
Issues about this test specifically: #26678
Failed: [k8s.io] ResourceQuota should verify ResourceQuota with best effort scope. {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/resource_quota.go:481
Expected error:
<*errors.StatusError | 0xc820d64100>: {
ErrStatus: {
TypeMeta: {Kind: "", APIVersion: ""},
ListMeta: {SelfLink: "", ResourceVersion: ""},
Status: "Failure",
Message: "the server does not allow access to the requested resource (get resourceQuotas quota-not-besteffort)",
Reason: "Forbidden",
Details: {
Name: "quota-not-besteffort",
Group: "",
Kind: "resourceQuotas",
Causes: [
{
Type: "UnexpectedServerResponse",
Message: "Forbidden: \"/api/v1/namespaces/e2e-tests-resourcequota-pjhoj/resourcequotas/quota-not-besteffort\"",
Field: "",
},
],
RetryAfterSeconds: 0,
},
Code: 403,
},
}
the server does not allow access to the requested resource (get resourceQuotas quota-not-besteffort)
not to have occurred
```
Failed: [k8s.io] Pods should not start app containers and fail the pod if init containers fail on a RestartNever pod {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pods.go:988
Jul 7 17:24:23.180: Error watching a pod: the server does not allow access to the requested resource (get pods)
```
Issues about this test specifically: #27465
Previous issues for this suite: #26742 #27839
|
1.0
|
kubernetes-e2e-gke: broken test run - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke/10735/
Multiple broken tests:
Failed: [k8s.io] Kubectl client [k8s.io] Update Demo should do a rolling update of a replication controller [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:233
Expected error:
<*errors.errorString | 0xc8207c7f70>: {
s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-f1wr3] [] 0xc820ae8ae0 Created update-demo-kitten\n Error from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)\n [] <nil> 0xc820ae9100 exit status 1 <nil> true [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba0a8 0xc8200ba130 0xc8200ba548] [0xa9ec00 0xa9ed60 0xa9ed60] 0xc820d6a6c0}:\nCommand stdout:\nCreated update-demo-kitten\n\nstderr:\nError from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)\n\nerror:\nexit status 1\n",
}
Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-f1wr3] [] 0xc820ae8ae0 Created update-demo-kitten
Error from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)
[] <nil> 0xc820ae9100 exit status 1 <nil> true [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba088 0xc8200ba540 0xc8200ba7a8] [0xc8200ba0a8 0xc8200ba130 0xc8200ba548] [0xa9ec00 0xa9ed60 0xa9ed60] 0xc820d6a6c0}:
Command stdout:
Created update-demo-kitten
stderr:
Error from server: the server does not allow access to the requested resource (put replicationControllers update-demo-nautilus)
error:
exit status 1
not to have occurred
```
Issues about this test specifically: #26425 #26715
Failed: [k8s.io] Kubectl client [k8s.io] Simple pod should support exec through an HTTP proxy {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:279
Expected error:
<*errors.errorString | 0xc8202aa550>: {
s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-4q19c] [] <nil> the server does not allow access to the requested resource (get services)\n [] <nil> 0xc82072a7c0 exit status 1 <nil> true [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081d0 0xc820108300] [0xa9ed60 0xa9ed60] 0xc820d062a0}:\nCommand stdout:\n\nstderr:\nthe server does not allow access to the requested resource (get services)\n\nerror:\nexit status 1\n",
}
Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.209.82 --kubeconfig=/workspace/.kube/config get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-4q19c] [] <nil> the server does not allow access to the requested resource (get services)
[] <nil> 0xc82072a7c0 exit status 1 <nil> true [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081b8 0xc8201081f0 0xc820108308] [0xc8201081d0 0xc820108300] [0xa9ed60 0xa9ed60] 0xc820d062a0}:
Command stdout:
stderr:
the server does not allow access to the requested resource (get services)
error:
exit status 1
not to have occurred
```
Issues about this test specifically: #27156
Failed: [k8s.io] Services should serve a basic endpoint from pods [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:133
Jul 7 17:24:27.653: Couldn't delete ns "e2e-tests-services-w86n1": the server does not allow access to the requested resource (delete namespaces e2e-tests-services-w86n1)
```
Issues about this test specifically: #26678
Failed: [k8s.io] ResourceQuota should verify ResourceQuota with best effort scope. {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/resource_quota.go:481
Expected error:
<*errors.StatusError | 0xc820d64100>: {
ErrStatus: {
TypeMeta: {Kind: "", APIVersion: ""},
ListMeta: {SelfLink: "", ResourceVersion: ""},
Status: "Failure",
Message: "the server does not allow access to the requested resource (get resourceQuotas quota-not-besteffort)",
Reason: "Forbidden",
Details: {
Name: "quota-not-besteffort",
Group: "",
Kind: "resourceQuotas",
Causes: [
{
Type: "UnexpectedServerResponse",
Message: "Forbidden: \"/api/v1/namespaces/e2e-tests-resourcequota-pjhoj/resourcequotas/quota-not-besteffort\"",
Field: "",
},
],
RetryAfterSeconds: 0,
},
Code: 403,
},
}
the server does not allow access to the requested resource (get resourceQuotas quota-not-besteffort)
not to have occurred
```
Failed: [k8s.io] Pods should not start app containers and fail the pod if init containers fail on a RestartNever pod {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pods.go:988
Jul 7 17:24:23.180: Error watching a pod: the server does not allow access to the requested resource (get pods)
```
Issues about this test specifically: #27465
Previous issues for this suite: #26742 #27839
|
test
|
kubernetes gke broken test run multiple broken tests failed kubectl client update demo should do a rolling update of a replication controller kubernetes suite go src io kubernetes output dockerized go src io kubernetes test kubectl go expected error s error running workspace kubernetes platforms linux kubectl created update demo kitten n error from server the server does not allow access to the requested resource put replicationcontrollers update demo nautilus n exit status true ncommand stdout ncreated update demo kitten n nstderr nerror from server the server does not allow access to the requested resource put replicationcontrollers update demo nautilus n nerror nexit status n error running workspace kubernetes platforms linux kubectl created update demo kitten error from server the server does not allow access to the requested resource put replicationcontrollers update demo nautilus exit status true command stdout created update demo kitten stderr error from server the server does not allow access to the requested resource put replicationcontrollers update demo nautilus error exit status not to have occurred issues about this test specifically failed kubectl client simple pod should support exec through an http proxy kubernetes suite go src io kubernetes output dockerized go src io kubernetes test kubectl go expected error s error running workspace kubernetes platforms linux kubectl the server does not allow access to the requested resource get services n exit status true ncommand stdout n nstderr nthe server does not allow access to the requested resource get services n nerror nexit status n error running workspace kubernetes platforms linux kubectl the server does not allow access to the requested resource get services exit status true command stdout stderr the server does not allow access to the requested resource get services error exit status not to have occurred issues about this test specifically failed services should serve a basic endpoint from pods kubernetes suite go src io kubernetes output dockerized go src io kubernetes test framework framework go jul couldn t delete ns tests services the server does not allow access to the requested resource delete namespaces tests services issues about this test specifically failed resourcequota should verify resourcequota with best effort scope kubernetes suite go src io kubernetes output dockerized go src io kubernetes test resource quota go expected error errstatus typemeta kind apiversion listmeta selflink resourceversion status failure message the server does not allow access to the requested resource get resourcequotas quota not besteffort reason forbidden details name quota not besteffort group kind resourcequotas causes type unexpectedserverresponse message forbidden api namespaces tests resourcequota pjhoj resourcequotas quota not besteffort field retryafterseconds code the server does not allow access to the requested resource get resourcequotas quota not besteffort not to have occurred failed pods should not start app containers and fail the pod if init containers fail on a restartnever pod kubernetes suite go src io kubernetes output dockerized go src io kubernetes test pods go jul error watching a pod the server does not allow access to the requested resource get pods issues about this test specifically previous issues for this suite
| 1
|
120,031
| 17,644,010,205
|
IssuesEvent
|
2021-08-20 01:27:18
|
AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
|
https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
|
opened
|
CVE-2021-37690 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2021-37690 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. In affected versions when running shape functions, some functions (such as `MutableHashTableShape`) produce extra output information in the form of a `ShapeAndType` struct. The shapes embedded in this struct are owned by an inference context that is cleaned up almost immediately; if the upstream code attempts to access this shape information, it can trigger a segfault. `ShapeRefiner` is mitigating this for normal output shapes by cloning them (and thus putting the newly created shape under ownership of an inference context that will not die), but we were not doing the same for shapes and types. This commit fixes that by doing similar logic on output shapes and types. We have patched the issue in GitHub commit ee119d4a498979525046fba1c3dd3f13a039fbb1. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-08-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37690>CVE-2021-37690</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-3hxh-8cp2-g4hg">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-3hxh-8cp2-g4hg</a></p>
<p>Release Date: 2021-08-13</p>
<p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-37690 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-37690 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. In affected versions when running shape functions, some functions (such as `MutableHashTableShape`) produce extra output information in the form of a `ShapeAndType` struct. The shapes embedded in this struct are owned by an inference context that is cleaned up almost immediately; if the upstream code attempts to access this shape information, it can trigger a segfault. `ShapeRefiner` is mitigating this for normal output shapes by cloning them (and thus putting the newly created shape under ownership of an inference context that will not die), but we were not doing the same for shapes and types. This commit fixes that by doing similar logic on output shapes and types. We have patched the issue in GitHub commit ee119d4a498979525046fba1c3dd3f13a039fbb1. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-08-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37690>CVE-2021-37690</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-3hxh-8cp2-g4hg">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-3hxh-8cp2-g4hg</a></p>
<p>Release Date: 2021-08-13</p>
<p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning in affected versions when running shape functions some functions such as mutablehashtableshape produce extra output information in the form of a shapeandtype struct the shapes embedded in this struct are owned by an inference context that is cleaned up almost immediately if the upstream code attempts to access this shape information it can trigger a segfault shaperefiner is mitigating this for normal output shapes by cloning them and thus putting the newly created shape under ownership of an inference context that will not die but we were not doing the same for shapes and types this commit fixes that by doing similar logic on output shapes and types we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
| 0
|
815,327
| 30,546,725,680
|
IssuesEvent
|
2023-07-20 04:59:02
|
Nexus-Mods/NexusMods.App
|
https://api.github.com/repos/Nexus-Mods/NexusMods.App
|
closed
|
Theme: Incorrect Value for StructuralBorderColor
|
complexity-small priority-low
|
# Bug Report
## Summary
In the theme, `StructuralBorderColor` is incorrectly defined as `#2DFFFFFF`, rather than `#303236`; which makes borders subtly different to their FIGMA designs. This difference is very subtle in most of the UI as the borders .
Current/App:

Intended/FIGMA:

## Other information
Already fixed as of 95d5b58964e3ee6f6bbb2e0f52c96b1b2b83de62. in branch cancel-downloads-new, this is just for progress tracking.
|
1.0
|
Theme: Incorrect Value for StructuralBorderColor - # Bug Report
## Summary
In the theme, `StructuralBorderColor` is incorrectly defined as `#2DFFFFFF`, rather than `#303236`; which makes borders subtly different to their FIGMA designs. This difference is very subtle in most of the UI as the borders .
Current/App:

Intended/FIGMA:

## Other information
Already fixed as of 95d5b58964e3ee6f6bbb2e0f52c96b1b2b83de62. in branch cancel-downloads-new, this is just for progress tracking.
|
non_test
|
theme incorrect value for structuralbordercolor bug report summary in the theme structuralbordercolor is incorrectly defined as rather than which makes borders subtly different to their figma designs this difference is very subtle in most of the ui as the borders current app intended figma other information already fixed as of in branch cancel downloads new this is just for progress tracking
| 0
|
676,671
| 23,133,724,825
|
IssuesEvent
|
2022-07-28 12:44:46
|
swlegion/tts
|
https://api.github.com/repos/swlegion/tts
|
opened
|
Solve 'The Darksaber' Problem
|
priority 2: soon 🐛 bug
|
Sabine and SC Maul both have an Upgrade called 'The Darksaber'. But the card has different art and rules. There are separate Armament entries for both, with requirements for the correct Units.
HOWEVER, when spawning the army you'll always get Sabine's Darksaber back because it is first in the array. So we need to check requirements if there are any.... or agree to special naming these and letting the list builders know the plan.
|
1.0
|
Solve 'The Darksaber' Problem - Sabine and SC Maul both have an Upgrade called 'The Darksaber'. But the card has different art and rules. There are separate Armament entries for both, with requirements for the correct Units.
HOWEVER, when spawning the army you'll always get Sabine's Darksaber back because it is first in the array. So we need to check requirements if there are any.... or agree to special naming these and letting the list builders know the plan.
|
non_test
|
solve the darksaber problem sabine and sc maul both have an upgrade called the darksaber but the card has different art and rules there are separate armament entries for both with requirements for the correct units however when spawning the army you ll always get sabine s darksaber back because it is first in the array so we need to check requirements if there are any or agree to special naming these and letting the list builders know the plan
| 0
|
336,715
| 24,510,951,887
|
IssuesEvent
|
2022-10-10 21:24:21
|
GeSoftColombia/sprintuno_iv
|
https://api.github.com/repos/GeSoftColombia/sprintuno_iv
|
closed
|
Creación de carpeta y archivo <home> para diligenciamiento de grupo
|
documentation Sprint_1
|
-[ ] Creación de carpeta y archivo para diligenciar grupo
|
1.0
|
Creación de carpeta y archivo <home> para diligenciamiento de grupo - -[ ] Creación de carpeta y archivo para diligenciar grupo
|
non_test
|
creación de carpeta y archivo para diligenciamiento de grupo creación de carpeta y archivo para diligenciar grupo
| 0
|
22,272
| 2,648,583,125
|
IssuesEvent
|
2015-03-14 01:59:40
|
dhamp/eiskaltdcpp
|
https://api.github.com/repos/dhamp/eiskaltdcpp
|
opened
|
Невозможность присоединения части пользователей без выделения слота
|
bug imported Priority-Medium
|
_From [yuri-xa...@yandex.ru](https://code.google.com/u/112318394159763967949/) on December 30, 2010 15:43:45_
Давненьео заметил такой дефект, но долго принимал за нехватку слотов просто...
Суть проблемы: многие пользователи не могут получить с меня информацию (список ли файлов или выложенный контент) без выделения им дополнительного слота посредством контекстного меню. При этом свободных слотов может быть с избытком.
_Original issue: http://code.google.com/p/eiskaltdc/issues/detail?id=923_
|
1.0
|
Невозможность присоединения части пользователей без выделения слота - _From [yuri-xa...@yandex.ru](https://code.google.com/u/112318394159763967949/) on December 30, 2010 15:43:45_
Давненьео заметил такой дефект, но долго принимал за нехватку слотов просто...
Суть проблемы: многие пользователи не могут получить с меня информацию (список ли файлов или выложенный контент) без выделения им дополнительного слота посредством контекстного меню. При этом свободных слотов может быть с избытком.
_Original issue: http://code.google.com/p/eiskaltdc/issues/detail?id=923_
|
non_test
|
невозможность присоединения части пользователей без выделения слота from on december давненьео заметил такой дефект но долго принимал за нехватку слотов просто суть проблемы многие пользователи не могут получить с меня информацию список ли файлов или выложенный контент без выделения им дополнительного слота посредством контекстного меню при этом свободных слотов может быть с избытком original issue
| 0
|
151,462
| 12,036,835,747
|
IssuesEvent
|
2020-04-13 20:35:01
|
ImagingDataCommons/IDC-WebApp
|
https://api.github.com/repos/ImagingDataCommons/IDC-WebApp
|
closed
|
BMI normal weight missing
|
bug testing needed testing passed
|
The BMI 'normal weight' is always 0. It looks like a Mysql/Solr mismatch. query_solr_and_format_result has the category 'normal' but in attr_by_source['related_set']['attributes']['bmi']['vals'] its 'normal weight'. I think we should go with 'normal weight'. Also why are counts so low? Was bmi just not recorded for most collections?
|
2.0
|
BMI normal weight missing - The BMI 'normal weight' is always 0. It looks like a Mysql/Solr mismatch. query_solr_and_format_result has the category 'normal' but in attr_by_source['related_set']['attributes']['bmi']['vals'] its 'normal weight'. I think we should go with 'normal weight'. Also why are counts so low? Was bmi just not recorded for most collections?
|
test
|
bmi normal weight missing the bmi normal weight is always it looks like a mysql solr mismatch query solr and format result has the category normal but in attr by source its normal weight i think we should go with normal weight also why are counts so low was bmi just not recorded for most collections
| 1
|
14,912
| 26,035,909,035
|
IssuesEvent
|
2022-12-22 04:56:33
|
seleniumbase/SeleniumBase
|
https://api.github.com/repos/seleniumbase/SeleniumBase
|
closed
|
Remove "options.headless" usage before SeleniumHQ deprecates it
|
requirements SeleniumBase 4
|
### Remove ``options.headless`` usage before SeleniumHQ deprecates it.
This is regarding: https://github.com/SeleniumHQ/selenium/issues/11467
Instead of setting ``options.headless``,
use ``options.add_argument("--headless")``
or ``options.add_argument("--headless=chrome")``.
(For Chrome and Edge)
This is mostly done. There were a few cases where both ways were being used together to set headless mode.
|
1.0
|
Remove "options.headless" usage before SeleniumHQ deprecates it - ### Remove ``options.headless`` usage before SeleniumHQ deprecates it.
This is regarding: https://github.com/SeleniumHQ/selenium/issues/11467
Instead of setting ``options.headless``,
use ``options.add_argument("--headless")``
or ``options.add_argument("--headless=chrome")``.
(For Chrome and Edge)
This is mostly done. There were a few cases where both ways were being used together to set headless mode.
|
non_test
|
remove options headless usage before seleniumhq deprecates it remove options headless usage before seleniumhq deprecates it this is regarding instead of setting options headless use options add argument headless or options add argument headless chrome for chrome and edge this is mostly done there were a few cases where both ways were being used together to set headless mode
| 0
|
3,953
| 2,543,832,470
|
IssuesEvent
|
2015-01-29 02:23:51
|
coollog/sublite
|
https://api.github.com/repos/coollog/sublite
|
closed
|
Add list of predefined industries
|
High Priority Long
|
- [x] add list of industries to company profile creation - @xtonyjiang
- [x] add list of industries to search - @coollog
- [x] change all current listings' industries to ones on the list
- [x] add option to select multiple industries for a company profile
|
1.0
|
Add list of predefined industries - - [x] add list of industries to company profile creation - @xtonyjiang
- [x] add list of industries to search - @coollog
- [x] change all current listings' industries to ones on the list
- [x] add option to select multiple industries for a company profile
|
non_test
|
add list of predefined industries add list of industries to company profile creation xtonyjiang add list of industries to search coollog change all current listings industries to ones on the list add option to select multiple industries for a company profile
| 0
|
6,909
| 2,824,171,912
|
IssuesEvent
|
2015-05-21 13:26:17
|
Code4HR/hrt-bus-finder
|
https://api.github.com/repos/Code4HR/hrt-bus-finder
|
opened
|
Reduce container width for larger screens
|
beginner design help wanted
|

Bootstrap by will increase the size of the container as the screen size increases. Override the Bootstrap CSS and set the maximum container size for the scheduling information to something like 540px.
```
.app-container {
max-width: 540px
}
```
|
1.0
|
Reduce container width for larger screens - 
Bootstrap by will increase the size of the container as the screen size increases. Override the Bootstrap CSS and set the maximum container size for the scheduling information to something like 540px.
```
.app-container {
max-width: 540px
}
```
|
non_test
|
reduce container width for larger screens bootstrap by will increase the size of the container as the screen size increases override the bootstrap css and set the maximum container size for the scheduling information to something like app container max width
| 0
|
529,201
| 15,383,250,277
|
IssuesEvent
|
2021-03-03 02:22:04
|
PyTorchLightning/pytorch-lightning
|
https://api.github.com/repos/PyTorchLightning/pytorch-lightning
|
closed
|
reproduciblity compared with vanilla pytorch
|
Priority P2 bug / fix good first issue help wanted won't fix
|
## 🐛 Bug
I am trying to compare a vanilla pytorch code which I refactored as a part of my learning pytorch-lightning but I see that the `training_step` is iterating through a different order of my dataset inspite of me setting `seed_everything` and `deterministic=True` and `benchmark=False`
## Please reproduce using [the BoringModel and post here]
I have this as the base, common for both vanilla mode and pytorch-lightning mode:
```
train_loader = torch.utils.data.DataLoader(\
MyDataSet(cfg=cfg, mode="train"), \
batch_size=16, \
shuffle=True)
```
In lightning I do:
```
def training_step(self, batch, batch_idx):
x, y, paths = batch
if Pathlib("check_input_path_pl").is_file():
append_write = 'a+' # append if already exists
else:
append_write = 'w' # make a new file if not
with open("check_input_path_pl", append_write) as f:
f.write(str(batch_idx) + " ==> " +str(paths[0]) + "\n")
y_hat = self(x)
loss = self.loss_func(y_hat, y)
return loss
```
in vanilla pytorch mode I do:
```
for batch_idx, (data, target, paths) in enumerate(train_loader):
data, target = data.to(device), target.to(device)
if Pathlib("check_input_path").is_file():
append_write = 'a+' # append if already exists
else:
append_write = 'w' # make a new file if not
with open("check_input_path", append_write) as f:
f.write(str(batch_idx) + " ==> " +str(paths[0]) + "\n")
```
but when I compare `check_input_path` with `check_input_path_pl` I see that after the first epoch, the order of the input file paths is different :(
I do:
```
seed_everything(111)
torch.backends.cudnn.deterministic = True
torch.backends.cudnn.benchmark = False
```
in both vanilla and lightning mode.
### Expected behavior
both should yield same order. To make a simple check, I iterate through the `train_loader` as soon as I create it in both pytl and pyt like so:
```
for epoch in range(0, 2):
for batch_idx, (data, target, paths) in enumerate(train_loader):
if Pathlib("paths2load").is_file():
append_write = 'a+' # append if already exists
else:
append_write = 'w' # make a new file if not
with open("paths2load", append_write) as f:
f.write(str(batch_idx) + " ==> " +str(paths[0]) + "\n")
```
and in both pyt and pytl the files show the same order. yet, when training, they load in different order. What could the reason be?
### Environment
pytorch-lightning==1.0.6
torch==1.6.0
torchvision==0.7.0
Debian Linux
PS: both pyt and pytl are running with the same virtualenv
|
1.0
|
reproduciblity compared with vanilla pytorch - ## 🐛 Bug
I am trying to compare a vanilla pytorch code which I refactored as a part of my learning pytorch-lightning but I see that the `training_step` is iterating through a different order of my dataset inspite of me setting `seed_everything` and `deterministic=True` and `benchmark=False`
## Please reproduce using [the BoringModel and post here]
I have this as the base, common for both vanilla mode and pytorch-lightning mode:
```
train_loader = torch.utils.data.DataLoader(\
MyDataSet(cfg=cfg, mode="train"), \
batch_size=16, \
shuffle=True)
```
In lightning I do:
```
def training_step(self, batch, batch_idx):
x, y, paths = batch
if Pathlib("check_input_path_pl").is_file():
append_write = 'a+' # append if already exists
else:
append_write = 'w' # make a new file if not
with open("check_input_path_pl", append_write) as f:
f.write(str(batch_idx) + " ==> " +str(paths[0]) + "\n")
y_hat = self(x)
loss = self.loss_func(y_hat, y)
return loss
```
in vanilla pytorch mode I do:
```
for batch_idx, (data, target, paths) in enumerate(train_loader):
data, target = data.to(device), target.to(device)
if Pathlib("check_input_path").is_file():
append_write = 'a+' # append if already exists
else:
append_write = 'w' # make a new file if not
with open("check_input_path", append_write) as f:
f.write(str(batch_idx) + " ==> " +str(paths[0]) + "\n")
```
but when I compare `check_input_path` with `check_input_path_pl` I see that after the first epoch, the order of the input file paths is different :(
I do:
```
seed_everything(111)
torch.backends.cudnn.deterministic = True
torch.backends.cudnn.benchmark = False
```
in both vanilla and lightning mode.
### Expected behavior
both should yield same order. To make a simple check, I iterate through the `train_loader` as soon as I create it in both pytl and pyt like so:
```
for epoch in range(0, 2):
for batch_idx, (data, target, paths) in enumerate(train_loader):
if Pathlib("paths2load").is_file():
append_write = 'a+' # append if already exists
else:
append_write = 'w' # make a new file if not
with open("paths2load", append_write) as f:
f.write(str(batch_idx) + " ==> " +str(paths[0]) + "\n")
```
and in both pyt and pytl the files show the same order. yet, when training, they load in different order. What could the reason be?
### Environment
pytorch-lightning==1.0.6
torch==1.6.0
torchvision==0.7.0
Debian Linux
PS: both pyt and pytl are running with the same virtualenv
|
non_test
|
reproduciblity compared with vanilla pytorch 🐛 bug i am trying to compare a vanilla pytorch code which i refactored as a part of my learning pytorch lightning but i see that the training step is iterating through a different order of my dataset inspite of me setting seed everything and deterministic true and benchmark false please reproduce using i have this as the base common for both vanilla mode and pytorch lightning mode train loader torch utils data dataloader mydataset cfg cfg mode train batch size shuffle true in lightning i do def training step self batch batch idx x y paths batch if pathlib check input path pl is file append write a append if already exists else append write w make a new file if not with open check input path pl append write as f f write str batch idx str paths n y hat self x loss self loss func y hat y return loss in vanilla pytorch mode i do for batch idx data target paths in enumerate train loader data target data to device target to device if pathlib check input path is file append write a append if already exists else append write w make a new file if not with open check input path append write as f f write str batch idx str paths n but when i compare check input path with check input path pl i see that after the first epoch the order of the input file paths is different i do seed everything torch backends cudnn deterministic true torch backends cudnn benchmark false in both vanilla and lightning mode expected behavior both should yield same order to make a simple check i iterate through the train loader as soon as i create it in both pytl and pyt like so for epoch in range for batch idx data target paths in enumerate train loader if pathlib is file append write a append if already exists else append write w make a new file if not with open append write as f f write str batch idx str paths n and in both pyt and pytl the files show the same order yet when training they load in different order what could the reason be environment pytorch lightning torch torchvision debian linux ps both pyt and pytl are running with the same virtualenv
| 0
|
350,925
| 31,932,538,284
|
IssuesEvent
|
2023-09-19 08:24:30
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix creation.test_triu_indices
|
Sub Task Ivy API Experimental Failing Test
|
| | |
|---|---|
|tensorflow|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix creation.test_triu_indices - | | |
|---|---|
|tensorflow|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
|
test
|
fix creation test triu indices tensorflow img src torch img src jax img src numpy img src paddle img src
| 1
|
20,750
| 10,915,110,557
|
IssuesEvent
|
2019-11-21 10:30:20
|
datalad/datalad
|
https://api.github.com/repos/datalad/datalad
|
opened
|
normalize_paths() can be a major slowdown for large N datasets
|
performance
|
Whenever an absolute path is provided, it will result in a `realpath()` call. This can be slow, depending on FS performance and path complexity. When done 500k times, it will have an impact.
|
True
|
normalize_paths() can be a major slowdown for large N datasets - Whenever an absolute path is provided, it will result in a `realpath()` call. This can be slow, depending on FS performance and path complexity. When done 500k times, it will have an impact.
|
non_test
|
normalize paths can be a major slowdown for large n datasets whenever an absolute path is provided it will result in a realpath call this can be slow depending on fs performance and path complexity when done times it will have an impact
| 0
|
29,927
| 7,134,887,558
|
IssuesEvent
|
2018-01-22 22:27:36
|
aimalz/chippr
|
https://api.github.com/repos/aimalz/chippr
|
closed
|
Replace basic probability distribution construction
|
Epic: code release Epic: null test
|
The whole `mvn.py`/`multi_dist.py`/`gmix.py`/`gauss.py`/`discrete.py` system is woefully inefficient. This issue can be closed when this redundant structure is eliminated in favor of something based on [`scipy.stats`](https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.rv_continuous.html) and/or [`qp`](https://github.com/aimalz/qp) and/or [`pomegranate`](https://github.com/jmschrei/pomegranate) objects that are much faster.
|
1.0
|
Replace basic probability distribution construction - The whole `mvn.py`/`multi_dist.py`/`gmix.py`/`gauss.py`/`discrete.py` system is woefully inefficient. This issue can be closed when this redundant structure is eliminated in favor of something based on [`scipy.stats`](https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.rv_continuous.html) and/or [`qp`](https://github.com/aimalz/qp) and/or [`pomegranate`](https://github.com/jmschrei/pomegranate) objects that are much faster.
|
non_test
|
replace basic probability distribution construction the whole mvn py multi dist py gmix py gauss py discrete py system is woefully inefficient this issue can be closed when this redundant structure is eliminated in favor of something based on and or and or objects that are much faster
| 0
|
381,201
| 11,274,881,289
|
IssuesEvent
|
2020-01-14 19:31:57
|
bcgov/entity
|
https://api.github.com/repos/bcgov/entity
|
opened
|
Page Validation Indicator in Stepper on File and Pay - UI DESIGN
|
ENTITY Priority2 UX
|
Include a page validation indicator in the stepper on "File and Pay".
The stepper allows users to view incomplete / invalid pages without selecting the "next step" button at the bottom of the page.
Therefore we can only perform an overall validation when the user selects "File and Pay" in the Confirm step.
When the user selects "File and Pay" an indicator would be displayed on any page that is not "valid".
When the user views a step with invalid data/fields the invalid fields are highlighted/indicated visually per the in-page validation.
|
1.0
|
Page Validation Indicator in Stepper on File and Pay - UI DESIGN - Include a page validation indicator in the stepper on "File and Pay".
The stepper allows users to view incomplete / invalid pages without selecting the "next step" button at the bottom of the page.
Therefore we can only perform an overall validation when the user selects "File and Pay" in the Confirm step.
When the user selects "File and Pay" an indicator would be displayed on any page that is not "valid".
When the user views a step with invalid data/fields the invalid fields are highlighted/indicated visually per the in-page validation.
|
non_test
|
page validation indicator in stepper on file and pay ui design include a page validation indicator in the stepper on file and pay the stepper allows users to view incomplete invalid pages without selecting the next step button at the bottom of the page therefore we can only perform an overall validation when the user selects file and pay in the confirm step when the user selects file and pay an indicator would be displayed on any page that is not valid when the user views a step with invalid data fields the invalid fields are highlighted indicated visually per the in page validation
| 0
|
229,061
| 18,279,503,582
|
IssuesEvent
|
2021-10-05 00:06:50
|
aces/Loris
|
https://api.github.com/repos/aces/Loris
|
opened
|
[Data Query Tool (Beta)] Saved queries show fields for private but not for public queries
|
Bug 24.0.0-testing
|
**Describe the bug**
Saved queries are displayed differently if saved privately or publicly. Private queries show all fields selected, whereas public queries do not.
**To Reproduce**
Steps to reproduce the behavior (attach screenshots if applicable):
1. Go to 'Data Query Tool (Beta)' module
2. Click on 'Load Existing Query' and compare 'User' and 'Shared' saved queries.
**What did you expect to happen?**
'User' and 'Shared' saved queries should be displayed in the same way.
**Browser Environment (please complete the following information):**
- Browser: Chrome
**Server Environment (if known):**
- LORIS Version: v24.0 testing

|
1.0
|
[Data Query Tool (Beta)] Saved queries show fields for private but not for public queries - **Describe the bug**
Saved queries are displayed differently if saved privately or publicly. Private queries show all fields selected, whereas public queries do not.
**To Reproduce**
Steps to reproduce the behavior (attach screenshots if applicable):
1. Go to 'Data Query Tool (Beta)' module
2. Click on 'Load Existing Query' and compare 'User' and 'Shared' saved queries.
**What did you expect to happen?**
'User' and 'Shared' saved queries should be displayed in the same way.
**Browser Environment (please complete the following information):**
- Browser: Chrome
**Server Environment (if known):**
- LORIS Version: v24.0 testing

|
test
|
saved queries show fields for private but not for public queries describe the bug saved queries are displayed differently if saved privately or publicly private queries show all fields selected whereas public queries do not to reproduce steps to reproduce the behavior attach screenshots if applicable go to data query tool beta module click on load existing query and compare user and shared saved queries what did you expect to happen user and shared saved queries should be displayed in the same way browser environment please complete the following information browser chrome server environment if known loris version testing
| 1
|
92,557
| 8,368,117,708
|
IssuesEvent
|
2018-10-04 14:02:26
|
italia/spid
|
https://api.github.com/repos/italia/spid
|
closed
|
Richiesta di Validazione Metadati - Regione Calabria
|
metadata nuovo md test
|
Per conto della Regione Calabria si richiede la verifica e il deploy dei metadati esposti all'URL:
https://ecosanita.regione.calabria.it/metadata-regionecalabria_20180912.xml
|
1.0
|
Richiesta di Validazione Metadati - Regione Calabria - Per conto della Regione Calabria si richiede la verifica e il deploy dei metadati esposti all'URL:
https://ecosanita.regione.calabria.it/metadata-regionecalabria_20180912.xml
|
test
|
richiesta di validazione metadati regione calabria per conto della regione calabria si richiede la verifica e il deploy dei metadati esposti all url
| 1
|
201,824
| 15,226,248,042
|
IssuesEvent
|
2021-02-18 08:37:24
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: ycsb/A/nodes=3 failed
|
C-test-failure O-roachtest O-robot branch-master release-blocker
|
[(roachtest).ycsb/A/nodes=3 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2688406&tab=buildLog) on [master@3c223f5f5162103110a790743b687ef2bf952489](https://github.com/cockroachdb/cockroach/commits/3c223f5f5162103110a790743b687ef2bf952489):
```
| 3.0s 0 1382.0 1030.3 0.9 1.1 1.8 8.9 read
| 3.0s 0 1380.0 1009.3 2.0 5.5 7.6 13.6 update
| 4.0s 0 2246.0 1334.2 0.9 1.3 2.4 6.6 read
| 4.0s 0 2186.0 1303.5 2.0 2.9 5.8 11.0 update
| 5.0s 0 2614.0 1590.2 0.9 1.4 2.4 6.8 read
| 5.0s 0 2668.0 1576.4 2.0 3.4 5.8 10.0 update
| 6.0s 0 3198.1 1858.2 0.9 1.5 2.6 58.7 read
| 6.0s 0 3223.1 1850.8 2.0 3.4 5.0 54.5 update
| 7.0s 0 3482.0 2090.1 0.9 1.6 2.9 6.8 read
| 7.0s 0 3627.0 2104.6 2.0 3.9 6.3 10.0 update
| 8.0s 0 3976.8 2326.0 0.9 2.0 3.4 8.4 read
| 8.0s 0 4005.8 2342.2 2.0 4.1 6.0 10.5 update
| 9.0s 0 4382.9 2554.5 1.0 2.0 3.5 8.4 read
| 9.0s 0 4304.9 2560.3 2.1 4.5 8.1 17.8 update
| 10.0s 0 4517.0 2750.8 1.0 2.2 3.7 9.4 read
| 10.0s 0 4536.0 2757.9 2.2 5.0 7.1 12.1 update
Wraps: (4) exit status 20
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError
cluster.go:2687,ycsb.go:62,ycsb.go:79,test_runner.go:767: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2675
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2683
| main.registerYCSB.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:62
| main.registerYCSB.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:79
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:767
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2731
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2645
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5652
| runtime.main
| /usr/local/go/src/runtime/proc.go:191
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<details><summary>More</summary><p>
Artifacts: [/ycsb/A/nodes=3](https://teamcity.cockroachdb.com/viewLog.html?buildId=2688406&tab=artifacts#/ycsb/A/nodes=3)
Related:
- #60369 roachtest: ycsb/A/nodes=3/cpu=32 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60368 roachtest: ycsb/A/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60088 roachtest: ycsb/A/nodes=3/cpu=32 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60087 roachtest: ycsb/A/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aycsb%2FA%2Fnodes%3D3.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: ycsb/A/nodes=3 failed - [(roachtest).ycsb/A/nodes=3 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2688406&tab=buildLog) on [master@3c223f5f5162103110a790743b687ef2bf952489](https://github.com/cockroachdb/cockroach/commits/3c223f5f5162103110a790743b687ef2bf952489):
```
| 3.0s 0 1382.0 1030.3 0.9 1.1 1.8 8.9 read
| 3.0s 0 1380.0 1009.3 2.0 5.5 7.6 13.6 update
| 4.0s 0 2246.0 1334.2 0.9 1.3 2.4 6.6 read
| 4.0s 0 2186.0 1303.5 2.0 2.9 5.8 11.0 update
| 5.0s 0 2614.0 1590.2 0.9 1.4 2.4 6.8 read
| 5.0s 0 2668.0 1576.4 2.0 3.4 5.8 10.0 update
| 6.0s 0 3198.1 1858.2 0.9 1.5 2.6 58.7 read
| 6.0s 0 3223.1 1850.8 2.0 3.4 5.0 54.5 update
| 7.0s 0 3482.0 2090.1 0.9 1.6 2.9 6.8 read
| 7.0s 0 3627.0 2104.6 2.0 3.9 6.3 10.0 update
| 8.0s 0 3976.8 2326.0 0.9 2.0 3.4 8.4 read
| 8.0s 0 4005.8 2342.2 2.0 4.1 6.0 10.5 update
| 9.0s 0 4382.9 2554.5 1.0 2.0 3.5 8.4 read
| 9.0s 0 4304.9 2560.3 2.1 4.5 8.1 17.8 update
| 10.0s 0 4517.0 2750.8 1.0 2.2 3.7 9.4 read
| 10.0s 0 4536.0 2757.9 2.2 5.0 7.1 12.1 update
Wraps: (4) exit status 20
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError
cluster.go:2687,ycsb.go:62,ycsb.go:79,test_runner.go:767: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2675
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2683
| main.registerYCSB.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:62
| main.registerYCSB.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:79
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:767
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2731
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2645
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5652
| runtime.main
| /usr/local/go/src/runtime/proc.go:191
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<details><summary>More</summary><p>
Artifacts: [/ycsb/A/nodes=3](https://teamcity.cockroachdb.com/viewLog.html?buildId=2688406&tab=artifacts#/ycsb/A/nodes=3)
Related:
- #60369 roachtest: ycsb/A/nodes=3/cpu=32 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60368 roachtest: ycsb/A/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60088 roachtest: ycsb/A/nodes=3/cpu=32 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60087 roachtest: ycsb/A/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aycsb%2FA%2Fnodes%3D3.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
roachtest ycsb a nodes failed on read update read update read update read update read update read update read update read update wraps exit status error types withstack withstack errutil withprefix main withcommanddetails exec exiterror cluster go ycsb go ycsb go test runner go monitor failure monitor task failed t fatal was called attached stack trace stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerycsb home agent work go src github com cockroachdb cockroach pkg cmd roachtest ycsb go main registerycsb home agent work go src github com cockroachdb cockroach pkg cmd roachtest ycsb go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps t fatal was called error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil leaferror more artifacts related roachtest ycsb a nodes cpu failed roachtest ycsb a nodes failed roachtest ycsb a nodes cpu failed roachtest ycsb a nodes failed powered by
| 1
|
204,173
| 15,421,627,551
|
IssuesEvent
|
2021-03-05 13:20:14
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
Reproducible Failure in org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites
|
:Distributed/Snapshot/Restore >test-failure Team:Distributed
|
Just ran into this locally working on the multiple page sizes cache but it reproduces on master as well:
```
./gradlew ':x-pack:plugin:searchable-snapshots:test' --tests "org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites" -Dtests.seed=163C909BF9F0F844 -Dtests.security.manager=true -Dtests.locale=sk-SK -Dtests.timezone=Etc/UTC -Druntime.java=15
```
fails with
```
[2021-03-04T20:58:27,650][INFO ][o.e.i.s.SearchableSnapshotDirectoryStatsTests] [testCachedBytesReadsAndWrites] before test
[2021-03-04T20:58:27,696][INFO ][o.e.e.NodeEnvironment ] [testCachedBytesReadsAndWrites] using [3] data paths, mounts [[/ (/dev/nvme0n1p2)]], net usable_space [121.6gb], net total_space [467.9gb], types [ext4]
[2021-03-04T20:58:27,697][INFO ][o.e.e.NodeEnvironment ] [testCachedBytesReadsAndWrites] heap size [15.6gb], compressed ordinary object pointers [true]
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by org.mockito.cglib.core.ReflectUtils$2 (file:/home/brownbear/.gradle/caches/modules-2/files-2.1/org.elasticsearch/securemock/1.2/98201d4ad5ac93f6b415ae9172d52b5e7cda490e/securemock-1.2.jar) to method java.lang.ClassLoader.defineClass(java.lang.String,byte[],int,int,java.security.ProtectionDomain)
WARNING: Please consider reporting this to the maintainers of org.mockito.cglib.core.ReflectUtils$2
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
[2021-03-04T20:58:27,922][INFO ][o.e.x.s.c.PersistentCache] [testCachedBytesReadsAndWrites] persistent cache index loaded
[2021-03-04T20:58:38,115][INFO ][o.e.i.s.SearchableSnapshotDirectoryStatsTests] [testCachedBytesReadsAndWrites] after test
REPRODUCE WITH: ./gradlew 'null' --tests "org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites" -Dtests.seed=163C909BF9F0F844 -Dtests.locale=sk-SK -Dtests.timezone=Etc/UTC -Druntime.java=15
java.lang.AssertionError:
Expected: <9641L>
but: was <9885L>
Expected :<9641L>
Actual :<9885L>
<Click to see difference>
at __randomizedtesting.SeedInfo.seed([163C909BF9F0F844:D38D21E9FBFE3FD4]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$2(SearchableSnapshotDirectoryStatsTests.java:131)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:955)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:928)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$3(SearchableSnapshotDirectoryStatsTests.java:129)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.executeTestCase(SearchableSnapshotDirectoryStatsTests.java:683)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.executeTestCaseWithCache(SearchableSnapshotDirectoryStatsTests.java:588)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites(SearchableSnapshotDirectoryStatsTests.java:116)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:64)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.base/java.lang.Thread.run(Thread.java:832)
Suppressed: java.lang.AssertionError:
Expected: <9641L>
but: was <9885L>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$2(SearchableSnapshotDirectoryStatsTests.java:131)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:943)
... 42 more
Suppressed: java.lang.AssertionError:
Expected: <9641L>
but: was <9885L>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$2(SearchableSnapshotDirectoryStatsTests.java:131)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:943)
... 42 more
Suppressed: java.lang.AssertionError:
```
on current master.
|
1.0
|
Reproducible Failure in org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites - Just ran into this locally working on the multiple page sizes cache but it reproduces on master as well:
```
./gradlew ':x-pack:plugin:searchable-snapshots:test' --tests "org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites" -Dtests.seed=163C909BF9F0F844 -Dtests.security.manager=true -Dtests.locale=sk-SK -Dtests.timezone=Etc/UTC -Druntime.java=15
```
fails with
```
[2021-03-04T20:58:27,650][INFO ][o.e.i.s.SearchableSnapshotDirectoryStatsTests] [testCachedBytesReadsAndWrites] before test
[2021-03-04T20:58:27,696][INFO ][o.e.e.NodeEnvironment ] [testCachedBytesReadsAndWrites] using [3] data paths, mounts [[/ (/dev/nvme0n1p2)]], net usable_space [121.6gb], net total_space [467.9gb], types [ext4]
[2021-03-04T20:58:27,697][INFO ][o.e.e.NodeEnvironment ] [testCachedBytesReadsAndWrites] heap size [15.6gb], compressed ordinary object pointers [true]
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by org.mockito.cglib.core.ReflectUtils$2 (file:/home/brownbear/.gradle/caches/modules-2/files-2.1/org.elasticsearch/securemock/1.2/98201d4ad5ac93f6b415ae9172d52b5e7cda490e/securemock-1.2.jar) to method java.lang.ClassLoader.defineClass(java.lang.String,byte[],int,int,java.security.ProtectionDomain)
WARNING: Please consider reporting this to the maintainers of org.mockito.cglib.core.ReflectUtils$2
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
[2021-03-04T20:58:27,922][INFO ][o.e.x.s.c.PersistentCache] [testCachedBytesReadsAndWrites] persistent cache index loaded
[2021-03-04T20:58:38,115][INFO ][o.e.i.s.SearchableSnapshotDirectoryStatsTests] [testCachedBytesReadsAndWrites] after test
REPRODUCE WITH: ./gradlew 'null' --tests "org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites" -Dtests.seed=163C909BF9F0F844 -Dtests.locale=sk-SK -Dtests.timezone=Etc/UTC -Druntime.java=15
java.lang.AssertionError:
Expected: <9641L>
but: was <9885L>
Expected :<9641L>
Actual :<9885L>
<Click to see difference>
at __randomizedtesting.SeedInfo.seed([163C909BF9F0F844:D38D21E9FBFE3FD4]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$2(SearchableSnapshotDirectoryStatsTests.java:131)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:955)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:928)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$3(SearchableSnapshotDirectoryStatsTests.java:129)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.executeTestCase(SearchableSnapshotDirectoryStatsTests.java:683)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.executeTestCaseWithCache(SearchableSnapshotDirectoryStatsTests.java:588)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.testCachedBytesReadsAndWrites(SearchableSnapshotDirectoryStatsTests.java:116)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:64)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.base/java.lang.Thread.run(Thread.java:832)
Suppressed: java.lang.AssertionError:
Expected: <9641L>
but: was <9885L>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$2(SearchableSnapshotDirectoryStatsTests.java:131)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:943)
... 42 more
Suppressed: java.lang.AssertionError:
Expected: <9641L>
but: was <9885L>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.index.store.SearchableSnapshotDirectoryStatsTests.lambda$testCachedBytesReadsAndWrites$2(SearchableSnapshotDirectoryStatsTests.java:131)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:943)
... 42 more
Suppressed: java.lang.AssertionError:
```
on current master.
|
test
|
reproducible failure in org elasticsearch index store searchablesnapshotdirectorystatstests testcachedbytesreadsandwrites just ran into this locally working on the multiple page sizes cache but it reproduces on master as well gradlew x pack plugin searchable snapshots test tests org elasticsearch index store searchablesnapshotdirectorystatstests testcachedbytesreadsandwrites dtests seed dtests security manager true dtests locale sk sk dtests timezone etc utc druntime java fails with before test using data paths mounts net usable space net total space types heap size compressed ordinary object pointers warning an illegal reflective access operation has occurred warning illegal reflective access by org mockito cglib core reflectutils file home brownbear gradle caches modules files org elasticsearch securemock securemock jar to method java lang classloader defineclass java lang string byte int int java security protectiondomain warning please consider reporting this to the maintainers of org mockito cglib core reflectutils warning use illegal access warn to enable warnings of further illegal reflective access operations warning all illegal access operations will be denied in a future release persistent cache index loaded after test reproduce with gradlew null tests org elasticsearch index store searchablesnapshotdirectorystatstests testcachedbytesreadsandwrites dtests seed dtests locale sk sk dtests timezone etc utc druntime java java lang assertionerror expected but was expected actual at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch index store searchablesnapshotdirectorystatstests lambda testcachedbytesreadsandwrites searchablesnapshotdirectorystatstests java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch index store searchablesnapshotdirectorystatstests lambda testcachedbytesreadsandwrites searchablesnapshotdirectorystatstests java at org elasticsearch index store searchablesnapshotdirectorystatstests executetestcase searchablesnapshotdirectorystatstests java at org elasticsearch index store searchablesnapshotdirectorystatstests executetestcasewithcache searchablesnapshotdirectorystatstests java at org elasticsearch index store searchablesnapshotdirectorystatstests testcachedbytesreadsandwrites searchablesnapshotdirectorystatstests java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java base java lang thread run thread java suppressed java lang assertionerror expected but was at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch index store searchablesnapshotdirectorystatstests lambda testcachedbytesreadsandwrites searchablesnapshotdirectorystatstests java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror expected but was at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch index store searchablesnapshotdirectorystatstests lambda testcachedbytesreadsandwrites searchablesnapshotdirectorystatstests java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror on current master
| 1
|
13,204
| 15,571,139,455
|
IssuesEvent
|
2021-03-17 04:12:01
|
Lothrazar/Cyclic
|
https://api.github.com/repos/Lothrazar/Cyclic
|
closed
|
Solidification Chamber not Processing
|
1.16 compatibility done
|
Minecraft Version: 1.16.5
Forge Version: 35.1.37
Mod Version: Cyclic-1.16.5-1.1.8.jar
Single Player or Server: Single Player
Describe problem (what you were doing; what happened; what should have happened):
I am currently playing the mod pack Sky Bee found on the curseforge application. When using the solidification chamber nothing happens. I have supplied the solidification chamber with 64000/64000 power and 46000/64000 honey to create things like honey apples, honey bottles, and crystalized amber. I supple the resources in the correct position and nothing is output from the machine as seen in the graphic below. I suspected it would be due to the version of the mod as sky bees hasn't updated to Cyclic-1.16.5-1.1.8.jar but I updated it to the latest version and the problem persists.
Log file link:
Video/images/gifs (direct upload or link):



|
True
|
Solidification Chamber not Processing - Minecraft Version: 1.16.5
Forge Version: 35.1.37
Mod Version: Cyclic-1.16.5-1.1.8.jar
Single Player or Server: Single Player
Describe problem (what you were doing; what happened; what should have happened):
I am currently playing the mod pack Sky Bee found on the curseforge application. When using the solidification chamber nothing happens. I have supplied the solidification chamber with 64000/64000 power and 46000/64000 honey to create things like honey apples, honey bottles, and crystalized amber. I supple the resources in the correct position and nothing is output from the machine as seen in the graphic below. I suspected it would be due to the version of the mod as sky bees hasn't updated to Cyclic-1.16.5-1.1.8.jar but I updated it to the latest version and the problem persists.
Log file link:
Video/images/gifs (direct upload or link):



|
non_test
|
solidification chamber not processing minecraft version forge version mod version cyclic jar single player or server single player describe problem what you were doing what happened what should have happened i am currently playing the mod pack sky bee found on the curseforge application when using the solidification chamber nothing happens i have supplied the solidification chamber with power and honey to create things like honey apples honey bottles and crystalized amber i supple the resources in the correct position and nothing is output from the machine as seen in the graphic below i suspected it would be due to the version of the mod as sky bees hasn t updated to cyclic jar but i updated it to the latest version and the problem persists log file link video images gifs direct upload or link
| 0
|
288,260
| 21,692,962,885
|
IssuesEvent
|
2022-05-09 17:05:49
|
withastro/docs
|
https://api.github.com/repos/withastro/docs
|
closed
|
Remove "experimental" from SSR page
|
improve documentation good first issue
|
This page needs updating now that SSR is no longer behind an experimental flag. It is the "new normal" but, it is still undergoing changes and in development. So, the text should at the very least remove reference to the experimental flag. But, perhaps keeping a note about it being in current development and subject to change is still helpful?
I suggest:
- Remove "(experimental)" from the SSR page heading at https://docs.astro.build/en/guides/server-side-rendering/
- remove "behind an experimental flag" to reflect that it is now enabled by default
- Maybe change "SSR is marked as experimental in Astro..." to something more like "SSR is new in Astro and changes will occur. before v1.0 stable release. Please keep up to date with API changes here." ?
|
1.0
|
Remove "experimental" from SSR page - This page needs updating now that SSR is no longer behind an experimental flag. It is the "new normal" but, it is still undergoing changes and in development. So, the text should at the very least remove reference to the experimental flag. But, perhaps keeping a note about it being in current development and subject to change is still helpful?
I suggest:
- Remove "(experimental)" from the SSR page heading at https://docs.astro.build/en/guides/server-side-rendering/
- remove "behind an experimental flag" to reflect that it is now enabled by default
- Maybe change "SSR is marked as experimental in Astro..." to something more like "SSR is new in Astro and changes will occur. before v1.0 stable release. Please keep up to date with API changes here." ?
|
non_test
|
remove experimental from ssr page this page needs updating now that ssr is no longer behind an experimental flag it is the new normal but it is still undergoing changes and in development so the text should at the very least remove reference to the experimental flag but perhaps keeping a note about it being in current development and subject to change is still helpful i suggest remove experimental from the ssr page heading at remove behind an experimental flag to reflect that it is now enabled by default maybe change ssr is marked as experimental in astro to something more like ssr is new in astro and changes will occur before stable release please keep up to date with api changes here
| 0
|
67,124
| 7,036,011,236
|
IssuesEvent
|
2017-12-28 05:34:09
|
Microsoft/vsts-tasks
|
https://api.github.com/repos/Microsoft/vsts-tasks
|
closed
|
VSTest: Publish fails if '[' or ']' are used in run title,
|
Area: Test
|
## Environment
- Server - VSTS or TFS on-premises?
VSTS
- Agent - Hosted or Private:
Private agent, windowsServer 2016, agent version: 2.126.0
## Issue Description
If the title of a test run contains '[' or ']' or both VSTest-Task will not publish the results and write the following warning: "Failed to publish test results: The given path's format is not supported.".
When i set system.debug to true the problem was obvious. The '##vso[results.publish'-command was cut and half the command was used as trx-path. See logs below:
### Error logs
[6_Test Assemblies test.dll;-obj.txt](https://github.com/Microsoft/vsts-tasks/files/1567219/6_Test.Assemblies.test.dll.-obj.txt)
Lines: 2242 -> 2250
|
1.0
|
VSTest: Publish fails if '[' or ']' are used in run title, - ## Environment
- Server - VSTS or TFS on-premises?
VSTS
- Agent - Hosted or Private:
Private agent, windowsServer 2016, agent version: 2.126.0
## Issue Description
If the title of a test run contains '[' or ']' or both VSTest-Task will not publish the results and write the following warning: "Failed to publish test results: The given path's format is not supported.".
When i set system.debug to true the problem was obvious. The '##vso[results.publish'-command was cut and half the command was used as trx-path. See logs below:
### Error logs
[6_Test Assemblies test.dll;-obj.txt](https://github.com/Microsoft/vsts-tasks/files/1567219/6_Test.Assemblies.test.dll.-obj.txt)
Lines: 2242 -> 2250
|
test
|
vstest publish fails if are used in run title environment server vsts or tfs on premises vsts agent hosted or private private agent windowsserver agent version issue description if the title of a test run contains or both vstest task will not publish the results and write the following warning failed to publish test results the given path s format is not supported when i set system debug to true the problem was obvious the vso results publish command was cut and half the command was used as trx path see logs below error logs lines
| 1
|
161,725
| 12,559,765,978
|
IssuesEvent
|
2020-06-07 19:58:34
|
valadaa48/retroroller
|
https://api.github.com/repos/valadaa48/retroroller
|
closed
|
es_systems.cfg needs platform set
|
please test
|
for these systems a platform needs set to scrape
saturn, famicom, super famicom, intellevision, master system
|
1.0
|
es_systems.cfg needs platform set - for these systems a platform needs set to scrape
saturn, famicom, super famicom, intellevision, master system
|
test
|
es systems cfg needs platform set for these systems a platform needs set to scrape saturn famicom super famicom intellevision master system
| 1
|
13,724
| 8,351,654,586
|
IssuesEvent
|
2018-10-02 01:31:14
|
azavea/tilegarden
|
https://api.github.com/repos/azavea/tilegarden
|
closed
|
Shrink aws-sdk
|
performance
|
Related to #117
We only really use the fetching-from-S3 part of `aws-sdk`, so it might be feasible to remove the parts of it that aren't used in order to slim down the bundle size. Rumor has it that it's available locally in the Lambda Node runtime, so check if that is true. If so, it might be possible to just relegate `aws-sdk` to a `devDependency`.
|
True
|
Shrink aws-sdk - Related to #117
We only really use the fetching-from-S3 part of `aws-sdk`, so it might be feasible to remove the parts of it that aren't used in order to slim down the bundle size. Rumor has it that it's available locally in the Lambda Node runtime, so check if that is true. If so, it might be possible to just relegate `aws-sdk` to a `devDependency`.
|
non_test
|
shrink aws sdk related to we only really use the fetching from part of aws sdk so it might be feasible to remove the parts of it that aren t used in order to slim down the bundle size rumor has it that it s available locally in the lambda node runtime so check if that is true if so it might be possible to just relegate aws sdk to a devdependency
| 0
|
155,225
| 12,244,202,990
|
IssuesEvent
|
2020-05-05 10:40:38
|
WoWManiaUK/Redemption
|
https://api.github.com/repos/WoWManiaUK/Redemption
|
closed
|
[QUEST] Where Dragons Fell-Killable Lich King
|
Fix - Tester Confirmed
|
**Links:**
https://www.wowhead.com/quest=13398/where-dragons-fell
**What is Happening:**
After turning the quest, one is shown the vision of Lich King resurrecting Sindragossa. The only problem being that Arthas is attackable and killable (28Million HP). He stands still while being damaged, does not attack, nor move, doesn't drop anything and doesn't provide one with any achievement.
**What Should happen:**
Vision should end, both Sindragosa and Arthas shouldn't be attackable, and both should disappear.
|
1.0
|
[QUEST] Where Dragons Fell-Killable Lich King - **Links:**
https://www.wowhead.com/quest=13398/where-dragons-fell
**What is Happening:**
After turning the quest, one is shown the vision of Lich King resurrecting Sindragossa. The only problem being that Arthas is attackable and killable (28Million HP). He stands still while being damaged, does not attack, nor move, doesn't drop anything and doesn't provide one with any achievement.
**What Should happen:**
Vision should end, both Sindragosa and Arthas shouldn't be attackable, and both should disappear.
|
test
|
where dragons fell killable lich king links what is happening after turning the quest one is shown the vision of lich king resurrecting sindragossa the only problem being that arthas is attackable and killable hp he stands still while being damaged does not attack nor move doesn t drop anything and doesn t provide one with any achievement what should happen vision should end both sindragosa and arthas shouldn t be attackable and both should disappear
| 1
|
9,333
| 3,036,654,985
|
IssuesEvent
|
2015-08-06 13:16:56
|
trendwerk/trendpress
|
https://api.github.com/repos/trendwerk/trendpress
|
closed
|
Front-end JS dependencies
|
feature needs-testing
|
Right now we don't use a dependency manager for front-end dependencies. The source is simply included within our repository. This makes repo's heavier than they should be.
[Bower](http://bower.io) seems the perfect solution to our problems. So far it seems like this has become the de facto standard.
|
1.0
|
Front-end JS dependencies - Right now we don't use a dependency manager for front-end dependencies. The source is simply included within our repository. This makes repo's heavier than they should be.
[Bower](http://bower.io) seems the perfect solution to our problems. So far it seems like this has become the de facto standard.
|
test
|
front end js dependencies right now we don t use a dependency manager for front end dependencies the source is simply included within our repository this makes repo s heavier than they should be seems the perfect solution to our problems so far it seems like this has become the de facto standard
| 1
|
177,573
| 13,731,390,918
|
IssuesEvent
|
2020-10-05 00:49:28
|
QubesOS/updates-status
|
https://api.github.com/repos/QubesOS/updates-status
|
closed
|
vmm-xen v4.14.0-3 (r4.1)
|
r4.1-bullseye-cur-test r4.1-buster-cur-test r4.1-centos8-cur-test r4.1-dom0-cur-test r4.1-stretch-cur-test
|
Update of vmm-xen to v4.14.0-3 for Qubes r4.1, see comments below for details.
Built from: https://github.com/QubesOS/qubes-vmm-xen/commit/6c426f6d7fd4630f82ac3aa75667572f364fcfa4
[Changes since previous version](https://github.com/QubesOS/qubes-vmm-xen/compare/v4.13.1-4...v4.14.0-3):
QubesOS/qubes-vmm-xen@6c426f6 version 4.14.0-3
QubesOS/qubes-vmm-xen@3caeccc debian: fix in-vm assumed xen-utils package version
QubesOS/qubes-vmm-xen@6e03e0d version 4.14.0-2
QubesOS/qubes-vmm-xen@cefb118 rpm: include xenhypfs tool
QubesOS/qubes-vmm-xen@62c3b97 rpm: fix qemu-xen packaging
QubesOS/qubes-vmm-xen@56591c0 rpm: remove remaining pygrub part
QubesOS/qubes-vmm-xen@853d1dd rpm: skip stubdom build at configure stage too
QubesOS/qubes-vmm-xen@13060f3 Fix seabios path
QubesOS/qubes-vmm-xen@1ad23b8 Add gcc warnings fixes to Debian package too
QubesOS/qubes-vmm-xen@343c2ae version 4.14.0-1
QubesOS/qubes-vmm-xen@cf5414b Merge remote-tracking branch 'origin/pr/83' into xen-4.14
QubesOS/qubes-vmm-xen@5f89d2d Adjust "Fix IGD passthrough with linux stubdomain" for Xen 4.14
QubesOS/qubes-vmm-xen@a1d2fd7 Merge remote-tracking branch 'origin/pr/79' into xen-4.14
QubesOS/qubes-vmm-xen@6370714 rpm: adjust dependencies for stubdomain API change
QubesOS/qubes-vmm-xen@f7f9a69 Update Debian and Arch packaging and patches for Xen 4.14
QubesOS/qubes-vmm-xen@7dfa12b Update to Xen 4.14.0
QubesOS/qubes-vmm-xen@4cabc08 spec: user python3_pkgversion macro
QubesOS/qubes-vmm-xen@f7f2427 Update travis
QubesOS/qubes-vmm-xen@08af9e0 IGD passthrough fix
Referenced issues:
QubesOS/qubes-issues#6050
QubesOS/qubes-issues#6029
If you're release manager, you can issue GPG-inline signed command:
* `Upload vmm-xen 6c426f6d7fd4630f82ac3aa75667572f364fcfa4 r4.1 current repo` (available 7 days from now)
* `Upload vmm-xen 6c426f6d7fd4630f82ac3aa75667572f364fcfa4 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload vmm-xen 6c426f6d7fd4630f82ac3aa75667572f364fcfa4 r4.1 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
|
5.0
|
vmm-xen v4.14.0-3 (r4.1) - Update of vmm-xen to v4.14.0-3 for Qubes r4.1, see comments below for details.
Built from: https://github.com/QubesOS/qubes-vmm-xen/commit/6c426f6d7fd4630f82ac3aa75667572f364fcfa4
[Changes since previous version](https://github.com/QubesOS/qubes-vmm-xen/compare/v4.13.1-4...v4.14.0-3):
QubesOS/qubes-vmm-xen@6c426f6 version 4.14.0-3
QubesOS/qubes-vmm-xen@3caeccc debian: fix in-vm assumed xen-utils package version
QubesOS/qubes-vmm-xen@6e03e0d version 4.14.0-2
QubesOS/qubes-vmm-xen@cefb118 rpm: include xenhypfs tool
QubesOS/qubes-vmm-xen@62c3b97 rpm: fix qemu-xen packaging
QubesOS/qubes-vmm-xen@56591c0 rpm: remove remaining pygrub part
QubesOS/qubes-vmm-xen@853d1dd rpm: skip stubdom build at configure stage too
QubesOS/qubes-vmm-xen@13060f3 Fix seabios path
QubesOS/qubes-vmm-xen@1ad23b8 Add gcc warnings fixes to Debian package too
QubesOS/qubes-vmm-xen@343c2ae version 4.14.0-1
QubesOS/qubes-vmm-xen@cf5414b Merge remote-tracking branch 'origin/pr/83' into xen-4.14
QubesOS/qubes-vmm-xen@5f89d2d Adjust "Fix IGD passthrough with linux stubdomain" for Xen 4.14
QubesOS/qubes-vmm-xen@a1d2fd7 Merge remote-tracking branch 'origin/pr/79' into xen-4.14
QubesOS/qubes-vmm-xen@6370714 rpm: adjust dependencies for stubdomain API change
QubesOS/qubes-vmm-xen@f7f9a69 Update Debian and Arch packaging and patches for Xen 4.14
QubesOS/qubes-vmm-xen@7dfa12b Update to Xen 4.14.0
QubesOS/qubes-vmm-xen@4cabc08 spec: user python3_pkgversion macro
QubesOS/qubes-vmm-xen@f7f2427 Update travis
QubesOS/qubes-vmm-xen@08af9e0 IGD passthrough fix
Referenced issues:
QubesOS/qubes-issues#6050
QubesOS/qubes-issues#6029
If you're release manager, you can issue GPG-inline signed command:
* `Upload vmm-xen 6c426f6d7fd4630f82ac3aa75667572f364fcfa4 r4.1 current repo` (available 7 days from now)
* `Upload vmm-xen 6c426f6d7fd4630f82ac3aa75667572f364fcfa4 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload vmm-xen 6c426f6d7fd4630f82ac3aa75667572f364fcfa4 r4.1 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
|
test
|
vmm xen update of vmm xen to for qubes see comments below for details built from qubesos qubes vmm xen version qubesos qubes vmm xen debian fix in vm assumed xen utils package version qubesos qubes vmm xen version qubesos qubes vmm xen rpm include xenhypfs tool qubesos qubes vmm xen rpm fix qemu xen packaging qubesos qubes vmm xen rpm remove remaining pygrub part qubesos qubes vmm xen rpm skip stubdom build at configure stage too qubesos qubes vmm xen fix seabios path qubesos qubes vmm xen add gcc warnings fixes to debian package too qubesos qubes vmm xen version qubesos qubes vmm xen merge remote tracking branch origin pr into xen qubesos qubes vmm xen adjust fix igd passthrough with linux stubdomain for xen qubesos qubes vmm xen merge remote tracking branch origin pr into xen qubesos qubes vmm xen rpm adjust dependencies for stubdomain api change qubesos qubes vmm xen update debian and arch packaging and patches for xen qubesos qubes vmm xen update to xen qubesos qubes vmm xen spec user pkgversion macro qubesos qubes vmm xen update travis qubesos qubes vmm xen igd passthrough fix referenced issues qubesos qubes issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload vmm xen current repo available days from now upload vmm xen current dists repo you can choose subset of distributions like vm vm available days from now upload vmm xen security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
| 1
|
116,471
| 11,913,787,437
|
IssuesEvent
|
2020-03-31 12:36:19
|
LEDApplications/lehd-schema
|
https://api.github.com/repos/LEDApplications/lehd-schema
|
opened
|
Change institution length
|
PSEO documentation
|
OPEID in past releases has been length 6, but we need to make it length 8 and have an updated OPEID list.
Things we need to do here:
- [ ] Update the CSV that feeds into schema.
- [ ] Change length of institution identifier.
|
1.0
|
Change institution length - OPEID in past releases has been length 6, but we need to make it length 8 and have an updated OPEID list.
Things we need to do here:
- [ ] Update the CSV that feeds into schema.
- [ ] Change length of institution identifier.
|
non_test
|
change institution length opeid in past releases has been length but we need to make it length and have an updated opeid list things we need to do here update the csv that feeds into schema change length of institution identifier
| 0
|
43,844
| 13,040,254,650
|
IssuesEvent
|
2020-07-28 18:10:44
|
LevyForchh/cadvisor
|
https://api.github.com/repos/LevyForchh/cadvisor
|
opened
|
CVE-2019-8331 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/google/cadvisor/stats-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/container/crio-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>bootstrap-4.0.0-beta.2.min.js</b>, <b>github.com/google/cadvisor/container/containerd-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/client/v2-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/info/v1-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>
<details><summary><b>github.com/google/cadvisor/stats-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/stats-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/container/crio-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/container/docker-33463ad2210c2490c2cfe822113ffe364d079eec
- github.com/google/cadvisor/container-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/container/crio-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- :x: **github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-4.0.0-beta.2.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0-beta.2/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0-beta.2/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /cadvisor/cmd/internal/pages/assets/js/bootstrap-4.0.0-beta.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-4.0.0-beta.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/container/containerd-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/container/docker-33463ad2210c2490c2cfe822113ffe364d079eec
- github.com/google/cadvisor/container-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/container/containerd-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/client/v2-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/integration/framework-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- :x: **github.com/google/cadvisor/client/v2-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/info/v1-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/utils/cpuload-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/utils/cpuload/netlink-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/info/v1-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/LevyForchh/cadvisor/commit/e4fb7b2d48835a6901c7b2e4a1bc7d1a57cab6b9">e4fb7b2d48835a6901c7b2e4a1bc7d1a57cab6b9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/google/cadvisor/stats","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/perf:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/stats:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/container/crio","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/docker:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/crio:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/perf","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/perf:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"4.0.0-beta.2","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:4.0.0-beta.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/container/containerd","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/docker:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/containerd:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/client/v2","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/integration/framework:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/client/v2:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/info/v1","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/utils/cpuload:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/utils/cpuload/netlink:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/info/v1:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"}],"vulnerabilityIdentifier":"CVE-2019-8331","vulnerabilityDetails":"In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-8331 (Medium) detected in multiple libraries - ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/google/cadvisor/stats-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/container/crio-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>bootstrap-4.0.0-beta.2.min.js</b>, <b>github.com/google/cadvisor/container/containerd-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/client/v2-33463ad2210c2490c2cfe822113ffe364d079eec</b>, <b>github.com/google/cadvisor/info/v1-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>
<details><summary><b>github.com/google/cadvisor/stats-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/stats-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/container/crio-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/container/docker-33463ad2210c2490c2cfe822113ffe364d079eec
- github.com/google/cadvisor/container-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/container/crio-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- :x: **github.com/google/cadvisor/perf-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-4.0.0-beta.2.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0-beta.2/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0-beta.2/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /cadvisor/cmd/internal/pages/assets/js/bootstrap-4.0.0-beta.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-4.0.0-beta.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/container/containerd-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/manager-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/container/docker-33463ad2210c2490c2cfe822113ffe364d079eec
- github.com/google/cadvisor/container-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/container/containerd-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/client/v2-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/integration/framework-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- :x: **github.com/google/cadvisor/client/v2-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<details><summary><b>github.com/google/cadvisor/info/v1-33463ad2210c2490c2cfe822113ffe364d079eec</b></p></summary>
<p>Analyzes resource usage and performance characteristics of running containers.</p>
<p>
Dependency Hierarchy:
- github.com/google/cadvisor/utils/cpuload-33463ad2210c2490c2cfe822113ffe364d079eec (Root Library)
- github.com/google/cadvisor/utils/cpuload/netlink-33463ad2210c2490c2cfe822113ffe364d079eec
- :x: **github.com/google/cadvisor/info/v1-33463ad2210c2490c2cfe822113ffe364d079eec** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/LevyForchh/cadvisor/commit/e4fb7b2d48835a6901c7b2e4a1bc7d1a57cab6b9">e4fb7b2d48835a6901c7b2e4a1bc7d1a57cab6b9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/google/cadvisor/stats","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/perf:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/stats:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/container/crio","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/docker:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/crio:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/perf","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/perf:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"4.0.0-beta.2","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:4.0.0-beta.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/container/containerd","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/manager:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/docker:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/container/containerd:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/client/v2","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/integration/framework:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/client/v2:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"},{"packageType":"GO","packageName":"github.com/google/cadvisor/info/v1","packageVersion":"33463ad2210c2490c2cfe822113ffe364d079eec","isTransitiveDependency":true,"dependencyTree":"github.com/google/cadvisor/utils/cpuload:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/utils/cpuload/netlink:33463ad2210c2490c2cfe822113ffe364d079eec;github.com/google/cadvisor/info/v1:33463ad2210c2490c2cfe822113ffe364d079eec","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1"}],"vulnerabilityIdentifier":"CVE-2019-8331","vulnerabilityDetails":"In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries github com google cadvisor stats github com google cadvisor container crio github com google cadvisor perf bootstrap beta min js github com google cadvisor container containerd github com google cadvisor client github com google cadvisor info github com google cadvisor stats analyzes resource usage and performance characteristics of running containers dependency hierarchy github com google cadvisor manager root library github com google cadvisor perf x github com google cadvisor stats vulnerable library github com google cadvisor container crio analyzes resource usage and performance characteristics of running containers dependency hierarchy github com google cadvisor manager root library github com google cadvisor container docker github com google cadvisor container x github com google cadvisor container crio vulnerable library github com google cadvisor perf analyzes resource usage and performance characteristics of running containers dependency hierarchy github com google cadvisor manager root library x github com google cadvisor perf vulnerable library bootstrap beta min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library cadvisor cmd internal pages assets js bootstrap beta min js dependency hierarchy x bootstrap beta min js vulnerable library github com google cadvisor container containerd analyzes resource usage and performance characteristics of running containers dependency hierarchy github com google cadvisor manager root library github com google cadvisor container docker github com google cadvisor container x github com google cadvisor container containerd vulnerable library github com google cadvisor client analyzes resource usage and performance characteristics of running containers dependency hierarchy github com google cadvisor integration framework root library x github com google cadvisor client vulnerable library github com google cadvisor info analyzes resource usage and performance characteristics of running containers dependency hierarchy github com google cadvisor utils cpuload root library github com google cadvisor utils cpuload netlink x github com google cadvisor info vulnerable library found in head commit a href vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap bootstrap sass isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before and x before xss is possible in the tooltip or popover data template attribute vulnerabilityurl
| 0
|
10,378
| 2,622,148,473
|
IssuesEvent
|
2015-03-04 00:05:07
|
byzhang/libcl
|
https://api.github.com/repos/byzhang/libcl
|
opened
|
The Test Crashed Under MY nv9600GT
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. after compiled and just running the TEST, with or without
debugging(VS2010Express),at lines:
65 // run tests
66 Log(INFO) << "****** calling radix sort ...";
67 testRadixSort(*lContext); // <-- video driver crash here
68 Log(INFO) << "****** done\n";
What is the expected output? What do you see instead?
the screen was totally black about 3-4 seconds, then it recovered with a
prompt, notified that the video drived is recovered from crashing... and any
console text later reported 'CL_OUT_OF_RESOURCES'
What version of the product are you using? On what operating system?
OS: Windows 7 32-bit
Adapter: NVIDIA 9600GT(512MB) with Driver 280.26, OpenCL 1.1 should be Profile
full supported(by AIDA 64)
Please provide any additional information below.
I tried it again the other computer with GT425M(1G), that's OK, Windows 7 but
64-bit, I guess it may be caused by the Driver but not sure.
See attachment for the screen output. Hope this is useful.
```
Original issue reported on code.google.com by `littlewa...@gmail.com` on 10 Sep 2011 at 12:35
Attachments:
* [output.txt](https://storage.googleapis.com/google-code-attachments/libcl/issue-1/comment-0/output.txt)
|
1.0
|
The Test Crashed Under MY nv9600GT - ```
What steps will reproduce the problem?
1. after compiled and just running the TEST, with or without
debugging(VS2010Express),at lines:
65 // run tests
66 Log(INFO) << "****** calling radix sort ...";
67 testRadixSort(*lContext); // <-- video driver crash here
68 Log(INFO) << "****** done\n";
What is the expected output? What do you see instead?
the screen was totally black about 3-4 seconds, then it recovered with a
prompt, notified that the video drived is recovered from crashing... and any
console text later reported 'CL_OUT_OF_RESOURCES'
What version of the product are you using? On what operating system?
OS: Windows 7 32-bit
Adapter: NVIDIA 9600GT(512MB) with Driver 280.26, OpenCL 1.1 should be Profile
full supported(by AIDA 64)
Please provide any additional information below.
I tried it again the other computer with GT425M(1G), that's OK, Windows 7 but
64-bit, I guess it may be caused by the Driver but not sure.
See attachment for the screen output. Hope this is useful.
```
Original issue reported on code.google.com by `littlewa...@gmail.com` on 10 Sep 2011 at 12:35
Attachments:
* [output.txt](https://storage.googleapis.com/google-code-attachments/libcl/issue-1/comment-0/output.txt)
|
non_test
|
the test crashed under my what steps will reproduce the problem after compiled and just running the test with or without debugging ,at lines run tests log info calling radix sort testradixsort lcontext video driver crash here log info done n what is the expected output what do you see instead the screen was totally black about seconds then it recovered with a prompt notified that the video drived is recovered from crashing and any console text later reported cl out of resources what version of the product are you using on what operating system os windows bit adapter nvidia with driver opencl should be profile full supported by aida please provide any additional information below i tried it again the other computer with that s ok windows but bit i guess it may be caused by the driver but not sure see attachment for the screen output hope this is useful original issue reported on code google com by littlewa gmail com on sep at attachments
| 0
|
7,538
| 10,687,846,394
|
IssuesEvent
|
2019-10-22 16:58:12
|
elheck/Squiddrone
|
https://api.github.com/repos/elheck/Squiddrone
|
opened
|
SW036 - SAFE LANDING, check for SYSTEMs underneath
|
Software Requirement
|
The SOFTWARE shall check for SYSTEMs underneath. The SOFTWARE shall use the GPS Positions of all other SYSTEMs and its sensors for this.
|
1.0
|
SW036 - SAFE LANDING, check for SYSTEMs underneath - The SOFTWARE shall check for SYSTEMs underneath. The SOFTWARE shall use the GPS Positions of all other SYSTEMs and its sensors for this.
|
non_test
|
safe landing check for systems underneath the software shall check for systems underneath the software shall use the gps positions of all other systems and its sensors for this
| 0
|
539,221
| 15,785,440,982
|
IssuesEvent
|
2021-04-01 16:20:32
|
zulip/zulip
|
https://api.github.com/repos/zulip/zulip
|
opened
|
Change model for messages with multiple image previews to stack images horizontally with scrollbar
|
area: message view help wanted priority: high
|
I'm not 100% sure this will be good UX, but it might be (and solves several problems), and we won't really know until we try.
Zulip's model for image previews is designed around a few goals:
* Not having images be so giant that they crowd out text (which is very common in other chat platforms with full-width image elements -- a square image can take ~50% of the screen)
* Avoiding the message feed jumping around randomly as images load (which requires that we can communicate to the browser via just HTML+CSS how much height the image element will consume before it is loaded). This can be quite problematic, since for scrolling convenience, we display in the DOM hundreds of messages of scrollback at a time, so it's very possible for a message well outside of view with 5 images to result in jumping scroll position.
* Avoiding creating extremely tall Zulip messages, which are not a great UX for keyboard users (one can scroll down partway through a message with e.g. `Space`, but overall if a message is multiple screens tall, it's a bit annoying).
The current design related to this is as follows:
* Image previews are 100px in height in CSS (And then can be whatever width that scales them to) -- this achieves the anti-crowding goal
* We show at most one image preview per line (this avoids the jumping -- since the browser doesn't need to know how the width of the images relates to the width of the message area and thus whether we'll end up wrapping to multiple lines).
* We show at most 5 image previews per message, to avoid creating extremely tall messages and also avoid unintended expensive behavior for rare automated bot messages (E.g. email gateway, etc.) that might have 100 links to images in them.
We've discussed options around this many times, but a new solution came up here: https://chat.zulip.org/#narrow/stream/92-learning/topic/Does.20using.20more.20than.205.20links.20remove.20image.20preview.3F/near/1148468
Specifically, we could just have all inline image previews going to the end of a message (or paragraph) be placed in a horizontal row together, with a horizontal scrollbar setup so that the browser knows that area will be 100px in height regardless of how many images we have there (avoiding the jumping bug).
And then we could raise the 5 to something like 10-15 (with maybe also a "show the first N" behavior, rather than "If more than N, show none") easily enough.
We'd need mobile responsive CSS to avoid an enormous horizontal scrolling region (for mobile web it should just be a vertical stack).
We can try doing this with just CSS, but it seems likely to me that we'll need to adjust the HTML markup to create a `div` to be the scrolling container for multiple images. I suppose we could try doing any HTML transformation in `rendered_markdown.js` rather than having the markdown processor generate the outer div; that might be better for prototyping since it doesn't adjust the database layout and would safer to test-deploy on chat.zulip.org to play with.
https://zulip.readthedocs.io/en/latest/subsystems/markdown.html has useful context.
|
1.0
|
Change model for messages with multiple image previews to stack images horizontally with scrollbar - I'm not 100% sure this will be good UX, but it might be (and solves several problems), and we won't really know until we try.
Zulip's model for image previews is designed around a few goals:
* Not having images be so giant that they crowd out text (which is very common in other chat platforms with full-width image elements -- a square image can take ~50% of the screen)
* Avoiding the message feed jumping around randomly as images load (which requires that we can communicate to the browser via just HTML+CSS how much height the image element will consume before it is loaded). This can be quite problematic, since for scrolling convenience, we display in the DOM hundreds of messages of scrollback at a time, so it's very possible for a message well outside of view with 5 images to result in jumping scroll position.
* Avoiding creating extremely tall Zulip messages, which are not a great UX for keyboard users (one can scroll down partway through a message with e.g. `Space`, but overall if a message is multiple screens tall, it's a bit annoying).
The current design related to this is as follows:
* Image previews are 100px in height in CSS (And then can be whatever width that scales them to) -- this achieves the anti-crowding goal
* We show at most one image preview per line (this avoids the jumping -- since the browser doesn't need to know how the width of the images relates to the width of the message area and thus whether we'll end up wrapping to multiple lines).
* We show at most 5 image previews per message, to avoid creating extremely tall messages and also avoid unintended expensive behavior for rare automated bot messages (E.g. email gateway, etc.) that might have 100 links to images in them.
We've discussed options around this many times, but a new solution came up here: https://chat.zulip.org/#narrow/stream/92-learning/topic/Does.20using.20more.20than.205.20links.20remove.20image.20preview.3F/near/1148468
Specifically, we could just have all inline image previews going to the end of a message (or paragraph) be placed in a horizontal row together, with a horizontal scrollbar setup so that the browser knows that area will be 100px in height regardless of how many images we have there (avoiding the jumping bug).
And then we could raise the 5 to something like 10-15 (with maybe also a "show the first N" behavior, rather than "If more than N, show none") easily enough.
We'd need mobile responsive CSS to avoid an enormous horizontal scrolling region (for mobile web it should just be a vertical stack).
We can try doing this with just CSS, but it seems likely to me that we'll need to adjust the HTML markup to create a `div` to be the scrolling container for multiple images. I suppose we could try doing any HTML transformation in `rendered_markdown.js` rather than having the markdown processor generate the outer div; that might be better for prototyping since it doesn't adjust the database layout and would safer to test-deploy on chat.zulip.org to play with.
https://zulip.readthedocs.io/en/latest/subsystems/markdown.html has useful context.
|
non_test
|
change model for messages with multiple image previews to stack images horizontally with scrollbar i m not sure this will be good ux but it might be and solves several problems and we won t really know until we try zulip s model for image previews is designed around a few goals not having images be so giant that they crowd out text which is very common in other chat platforms with full width image elements a square image can take of the screen avoiding the message feed jumping around randomly as images load which requires that we can communicate to the browser via just html css how much height the image element will consume before it is loaded this can be quite problematic since for scrolling convenience we display in the dom hundreds of messages of scrollback at a time so it s very possible for a message well outside of view with images to result in jumping scroll position avoiding creating extremely tall zulip messages which are not a great ux for keyboard users one can scroll down partway through a message with e g space but overall if a message is multiple screens tall it s a bit annoying the current design related to this is as follows image previews are in height in css and then can be whatever width that scales them to this achieves the anti crowding goal we show at most one image preview per line this avoids the jumping since the browser doesn t need to know how the width of the images relates to the width of the message area and thus whether we ll end up wrapping to multiple lines we show at most image previews per message to avoid creating extremely tall messages and also avoid unintended expensive behavior for rare automated bot messages e g email gateway etc that might have links to images in them we ve discussed options around this many times but a new solution came up here specifically we could just have all inline image previews going to the end of a message or paragraph be placed in a horizontal row together with a horizontal scrollbar setup so that the browser knows that area will be in height regardless of how many images we have there avoiding the jumping bug and then we could raise the to something like with maybe also a show the first n behavior rather than if more than n show none easily enough we d need mobile responsive css to avoid an enormous horizontal scrolling region for mobile web it should just be a vertical stack we can try doing this with just css but it seems likely to me that we ll need to adjust the html markup to create a div to be the scrolling container for multiple images i suppose we could try doing any html transformation in rendered markdown js rather than having the markdown processor generate the outer div that might be better for prototyping since it doesn t adjust the database layout and would safer to test deploy on chat zulip org to play with has useful context
| 0
|
723,146
| 24,886,458,376
|
IssuesEvent
|
2022-10-28 08:12:32
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
slideshow not color managed
|
priority: high difficulty: trivial bug: pending release notes: pending
|
All is said in the title. To reproduce:
- select a color space that is very visible (like linear rec 2020)
- enter the slideshow
- see how different the image is compared to lighttable or darkroom
|
1.0
|
slideshow not color managed - All is said in the title. To reproduce:
- select a color space that is very visible (like linear rec 2020)
- enter the slideshow
- see how different the image is compared to lighttable or darkroom
|
non_test
|
slideshow not color managed all is said in the title to reproduce select a color space that is very visible like linear rec enter the slideshow see how different the image is compared to lighttable or darkroom
| 0
|
7,696
| 10,863,660,514
|
IssuesEvent
|
2019-11-14 15:30:24
|
Shopify/kubernetes-deploy
|
https://api.github.com/repos/Shopify/kubernetes-deploy
|
closed
|
Write migration guide.
|
:rocket: 1.0 requirement
|
We need to write a guide that explains how to migrate from `exe/kubernetes-*` to `krane *`.
- Provide mappings for flags that have changed.
- Provide examples of how to deploy global resources
- Provide examples of any breaking changes
- Explain how to install k8s-deploy gem and krane gem.
|
1.0
|
Write migration guide. - We need to write a guide that explains how to migrate from `exe/kubernetes-*` to `krane *`.
- Provide mappings for flags that have changed.
- Provide examples of how to deploy global resources
- Provide examples of any breaking changes
- Explain how to install k8s-deploy gem and krane gem.
|
non_test
|
write migration guide we need to write a guide that explains how to migrate from exe kubernetes to krane provide mappings for flags that have changed provide examples of how to deploy global resources provide examples of any breaking changes explain how to install deploy gem and krane gem
| 0
|
779,950
| 27,373,222,048
|
IssuesEvent
|
2023-02-28 02:18:46
|
UNopenGIS/7
|
https://api.github.com/repos/UNopenGIS/7
|
closed
|
Releasable Basemap Tiles (RBT) を取り込みたい
|
priority/MAY
|
## See Also
https://github.com/agcgeoint/rbt
## 私ならこう取り込みたい
- MBTiles で配布されているようなので、PMTiles に変換する
- style.json を charites 経由で加工する
- MapLibre GL JS ベースでローカル完結したウェブ地図を作る
|
1.0
|
Releasable Basemap Tiles (RBT) を取り込みたい - ## See Also
https://github.com/agcgeoint/rbt
## 私ならこう取り込みたい
- MBTiles で配布されているようなので、PMTiles に変換する
- style.json を charites 経由で加工する
- MapLibre GL JS ベースでローカル完結したウェブ地図を作る
|
non_test
|
releasable basemap tiles rbt を取り込みたい see also 私ならこう取り込みたい mbtiles で配布されているようなので、pmtiles に変換する style json を charites 経由で加工する maplibre gl js ベースでローカル完結したウェブ地図を作る
| 0
|
439,748
| 12,686,326,393
|
IssuesEvent
|
2020-06-20 10:13:28
|
din0s/ActionHeroes
|
https://api.github.com/repos/din0s/ActionHeroes
|
closed
|
Make scroll-up button a global feature.
|
enhancement low priority
|
The scroll-up button is currently implemented on the Landing-Page only. It needs to be relocated so it can be a global feature, applied to all our pages.
|
1.0
|
Make scroll-up button a global feature. - The scroll-up button is currently implemented on the Landing-Page only. It needs to be relocated so it can be a global feature, applied to all our pages.
|
non_test
|
make scroll up button a global feature the scroll up button is currently implemented on the landing page only it needs to be relocated so it can be a global feature applied to all our pages
| 0
|
230,942
| 18,725,736,390
|
IssuesEvent
|
2021-11-03 16:05:56
|
flutter/devtools
|
https://api.github.com/repos/flutter/devtools
|
closed
|
service manager related test flakes
|
testing fix it friday
|
This same assertion error for `'isolate != null'` has been seen on other tests as well.
```
02:58 +241 ~3: /home/runner/work/devtools/devtools/packages/devtools_app/test/inspector_service_test.dart: inspector service tests track widget creation on
PreferencesController: storage not initialized
02:59 +241 ~3 -1: /home/runner/work/devtools/devtools/packages/devtools_app/test/inspector_service_test.dart: inspector service tests track widget creation on [E]
'package:devtools_app/src/service_manager.dart': Failed assertion: line 489 pos 12: 'isolate != null': is not true.
dart:core _AssertionError._throwNew
package:devtools_app/src/service_manager.dart 489:12 ServiceConnectionManager.libraryUriAvailableNow
package:devtools_app/src/service_manager.dart 499:12 ServiceConnectionManager.libraryUriAvailable
Retry: inspector service tests track widget creation on
'package:devtools_app/src/inspector/inspector_service.dart': Failed assertion: line 53 pos 16: 'serviceManager.connectedAppInitialized': is not true.
dart:core _AssertionError._throwNew
package:devtools_app/src/inspector/inspector_service.dart 53:16 new InspectorService
test/inspector_service_test.dart 26:24 main.<fn>
test/inspector_service_test.dart 25:25 main.<fn>
package:devtools_test/flutter_test_environment.dart 107:57 FlutterTestEnvironment.setupEnvironment
test/inspector_service_test.dart 44:19 main.<fn>.<fn>
test/inspector_service_test.dart 43:40 main.<fn>.<fn>
```
|
1.0
|
service manager related test flakes - This same assertion error for `'isolate != null'` has been seen on other tests as well.
```
02:58 +241 ~3: /home/runner/work/devtools/devtools/packages/devtools_app/test/inspector_service_test.dart: inspector service tests track widget creation on
PreferencesController: storage not initialized
02:59 +241 ~3 -1: /home/runner/work/devtools/devtools/packages/devtools_app/test/inspector_service_test.dart: inspector service tests track widget creation on [E]
'package:devtools_app/src/service_manager.dart': Failed assertion: line 489 pos 12: 'isolate != null': is not true.
dart:core _AssertionError._throwNew
package:devtools_app/src/service_manager.dart 489:12 ServiceConnectionManager.libraryUriAvailableNow
package:devtools_app/src/service_manager.dart 499:12 ServiceConnectionManager.libraryUriAvailable
Retry: inspector service tests track widget creation on
'package:devtools_app/src/inspector/inspector_service.dart': Failed assertion: line 53 pos 16: 'serviceManager.connectedAppInitialized': is not true.
dart:core _AssertionError._throwNew
package:devtools_app/src/inspector/inspector_service.dart 53:16 new InspectorService
test/inspector_service_test.dart 26:24 main.<fn>
test/inspector_service_test.dart 25:25 main.<fn>
package:devtools_test/flutter_test_environment.dart 107:57 FlutterTestEnvironment.setupEnvironment
test/inspector_service_test.dart 44:19 main.<fn>.<fn>
test/inspector_service_test.dart 43:40 main.<fn>.<fn>
```
|
test
|
service manager related test flakes this same assertion error for isolate null has been seen on other tests as well home runner work devtools devtools packages devtools app test inspector service test dart inspector service tests track widget creation on preferencescontroller storage not initialized home runner work devtools devtools packages devtools app test inspector service test dart inspector service tests track widget creation on package devtools app src service manager dart failed assertion line pos isolate null is not true dart core assertionerror thrownew package devtools app src service manager dart serviceconnectionmanager libraryuriavailablenow package devtools app src service manager dart serviceconnectionmanager libraryuriavailable retry inspector service tests track widget creation on package devtools app src inspector inspector service dart failed assertion line pos servicemanager connectedappinitialized is not true dart core assertionerror thrownew package devtools app src inspector inspector service dart new inspectorservice test inspector service test dart main test inspector service test dart main package devtools test flutter test environment dart fluttertestenvironment setupenvironment test inspector service test dart main test inspector service test dart main
| 1
|
32,923
| 15,716,155,249
|
IssuesEvent
|
2021-03-28 05:37:40
|
diofant/diofant
|
https://api.github.com/repos/diofant/diofant
|
opened
|
Slow gcd for multivariate polynomials over finite fields
|
enhancement performance polys
|
See e.g. example [here](https://github.com/diofant/diofant/pull/1132#issuecomment-803298312) and below (with an obvious replacement `s/from_sympy/from_expr/`). c.f. [rings](https://rings.readthedocs.io/en/latest/) library.
|
True
|
Slow gcd for multivariate polynomials over finite fields - See e.g. example [here](https://github.com/diofant/diofant/pull/1132#issuecomment-803298312) and below (with an obvious replacement `s/from_sympy/from_expr/`). c.f. [rings](https://rings.readthedocs.io/en/latest/) library.
|
non_test
|
slow gcd for multivariate polynomials over finite fields see e g example and below with an obvious replacement s from sympy from expr c f library
| 0
|
214,433
| 16,588,933,665
|
IssuesEvent
|
2021-06-01 04:15:52
|
thanos-io/thanos
|
https://api.github.com/repos/thanos-io/thanos
|
closed
|
Add E2E test for exemplars API
|
component: query difficulty: medium good first issue help wanted tests
|
Now the exemplars API is added and we already instrumented Thanos with exemplars in https://github.com/thanos-io/thanos/pull/3977. It is time to add an E2E test case for it.
AC:
~~1. Add a Jaeger all in one service in the E2E test as it is required for tracing~~
2. Update the options to start Thanos component with tracing configs
3. Call the exemplars API from thanos querier and check the results
For the request and response of the exemplars API, please check https://prometheus.io/docs/prometheus/latest/querying/api/#querying-exemplars
|
1.0
|
Add E2E test for exemplars API - Now the exemplars API is added and we already instrumented Thanos with exemplars in https://github.com/thanos-io/thanos/pull/3977. It is time to add an E2E test case for it.
AC:
~~1. Add a Jaeger all in one service in the E2E test as it is required for tracing~~
2. Update the options to start Thanos component with tracing configs
3. Call the exemplars API from thanos querier and check the results
For the request and response of the exemplars API, please check https://prometheus.io/docs/prometheus/latest/querying/api/#querying-exemplars
|
test
|
add test for exemplars api now the exemplars api is added and we already instrumented thanos with exemplars in it is time to add an test case for it ac add a jaeger all in one service in the test as it is required for tracing update the options to start thanos component with tracing configs call the exemplars api from thanos querier and check the results for the request and response of the exemplars api please check
| 1
|
629,030
| 20,021,507,823
|
IssuesEvent
|
2022-02-01 16:48:14
|
CLOSER-Cohorts/archivist
|
https://api.github.com/repos/CLOSER-Cohorts/archivist
|
opened
|
Security patches
|
High priority
|
There are a whole stack of security patches that should probably be applied before they get completely out of hand
|
1.0
|
Security patches - There are a whole stack of security patches that should probably be applied before they get completely out of hand
|
non_test
|
security patches there are a whole stack of security patches that should probably be applied before they get completely out of hand
| 0
|
484,113
| 13,934,598,585
|
IssuesEvent
|
2020-10-22 10:15:23
|
AY2021S1-CS2113T-F11-1/tp
|
https://api.github.com/repos/AY2021S1-CS2113T-F11-1/tp
|
closed
|
As a user, I would like to search my past records based on input
|
diet priority.High type.Story
|
so that I can find my record faster.
|
1.0
|
As a user, I would like to search my past records based on input - so that I can find my record faster.
|
non_test
|
as a user i would like to search my past records based on input so that i can find my record faster
| 0
|
41,057
| 5,294,467,802
|
IssuesEvent
|
2017-02-09 10:52:39
|
IMA-WorldHealth/bhima-2.X
|
https://api.github.com/repos/IMA-WorldHealth/bhima-2.X
|
closed
|
(design) Useful MySQL views for validation, quick lookups
|
design
|
A MySQL view is a stored query that looks like a table. See [this page](http://dev.mysql.com/doc/refman/5.7/en/views.html) for more information.
I propose that we create a few views that will help our validation queries:
1. `entity`
``` sql
CREATE VIEW entity AS
SELECT uuid, text FROM debtor
UNION
SELECT uuid, text FROM creditor;
```
2. `record`
``` sql
CREATE VIEW record AS
SELECT DISTINCT record_uuid, trans_id FROM posting_journal
UNION
SELECT DISTINCT record_uuid, trans_id FROM general_ledger;
```
These queries/view will be helpful when checking rapidly if a `reference_uuid` exists or a `entity_uuid` exists when verifying that data entered into the Posting Journal is correct (on the server).
|
1.0
|
(design) Useful MySQL views for validation, quick lookups - A MySQL view is a stored query that looks like a table. See [this page](http://dev.mysql.com/doc/refman/5.7/en/views.html) for more information.
I propose that we create a few views that will help our validation queries:
1. `entity`
``` sql
CREATE VIEW entity AS
SELECT uuid, text FROM debtor
UNION
SELECT uuid, text FROM creditor;
```
2. `record`
``` sql
CREATE VIEW record AS
SELECT DISTINCT record_uuid, trans_id FROM posting_journal
UNION
SELECT DISTINCT record_uuid, trans_id FROM general_ledger;
```
These queries/view will be helpful when checking rapidly if a `reference_uuid` exists or a `entity_uuid` exists when verifying that data entered into the Posting Journal is correct (on the server).
|
non_test
|
design useful mysql views for validation quick lookups a mysql view is a stored query that looks like a table see for more information i propose that we create a few views that will help our validation queries entity sql create view entity as select uuid text from debtor union select uuid text from creditor record sql create view record as select distinct record uuid trans id from posting journal union select distinct record uuid trans id from general ledger these queries view will be helpful when checking rapidly if a reference uuid exists or a entity uuid exists when verifying that data entered into the posting journal is correct on the server
| 0
|
50,317
| 6,354,221,720
|
IssuesEvent
|
2017-07-29 07:15:43
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
C# Design Notes for May 3-4, 2016
|
Area-Language Design Design Notes Language-C# Language-VB New Language Feature - Pattern Matching New Language Feature - Tuples
|
# C# Design Notes for May 3-4, 2016
This pair of meetings further explored the space around tuple syntax, pattern matching and deconstruction.
1. Deconstructors - how to specify them
2. Switch conversions - how to deal with them
3. Tuple conversions - how to do them
4. Tuple-like types - how to construct them
Lots of concrete decisions, that allow us to make progress on implementation.
# Deconstructors
In #11031 we discussed the different contexts in which deconstruction should be able to occur, namely deconstructing _assignment_ (into existing variables), _declaration_ (into freshly declared local variables) and _patterns_ (as part of applying a recursive pattern).
We also explored the design space of how exactly "deconstructability" should be specified for a given type, but left the decision open - until now. Here's what we decided - and why. We'll stick to these decisions in initial prototypes, but as always are willing to be swayed by evidence as we roll them out and get usage.
**_Deconstruction should be specified with an instance (or extension) method**_. This is in keeping with other API patterns added throughout the history of C#, such as `GetEnumerator`, `Add`, and `GetAwaiter`. The benefit is that this leads to a relatively natural kind of member to have, and it can be specified with an extension method so that existing types can be augmented to be deconstructable outside of their own code.
The choice limits the ability of the pattern to later grow up to facilitate "active patterns". We aren't too concerned about that, because if we want to add active patterns at a later date we can easily come up with a separate mechanism for specifying those.
**_The instance/extension method should be called `Deconstruct`**_. We've been informally calling it `GetValues` for a while, but that name suffers from being in too popular use already, and not always for a similar purpose. This is a decision we're willing to alter if a better name comes along, and is sufficiently unencumbered.
**_The Deconstruct method "returns" the component values by use of individual out parameters**_. This choice may seem odd: after all we're adding a perfectly great feature called tuples, just so that you can return multiple values! The motivation here is primarily that we want `Deconstruct` to be overloadable. Sometimes there are genuinely multiple ways to deconstruct, and sometimes the type evolves over time to add more properties, and as you extend the `Deconstruct` method you also want to leave an old overload available for source and binary compat.
This one does nag us a little, because the declaration form with tuples is so much simpler, and would be sufficient in a majority of cases. On the other hand, this allows us to declare decomposition logic _for_ tuples the same way as for other types, which we couldn't if we depended on tuples for it!
Should this become a major nuisance (we don't think so) one could consider a hybrid approach where both tuple-returning and out-parameter versions were recognized, but for now we won't.
All in all, the deconstructor pattern looks like one of these:
``` c#
class Name
{
public void Deconstruct(out string first, out string last) { first = First; last = Last; }
...
}
// or
static class Extensions
{
public static void Deconstruct(this Name name, out string first, out string last) { first = name.First; last = name.Last; }
}
```
# Switch conversions
Switch statements today have a wrinkle where they will apply a unique implicit conversion from the switched-on expression to a (currently) switchable type. As we expand to allow switching on any type, this may be confusing at times, but we need to keep it at least in some scenarios, for backwards compatibility.
``` c#
switch (expr) // of some type Expression, which "cleverly" has a user defined conversion to int for evaluation
{
case Constant(int i): ... // Won't work, though Constant derives from Expression, because expr has been converted to int
...
}
```
Our current stance is that this is fringe enough for us to ignore. If you run into such a conversion and didn't want it, you'll have to work around it, e.g. by casting your switch expression to object.
If this turns out to be more of a nuisance we may have to come up with a smarter rule, but for now we're good with this.
# Tuple conversions
In #11031 we decided to add tuple conversions, that essentially convert tuples whenever their elements convert - unlike the more restrictive conversions that follow from `ValueTuple<...>` being a generic struct. In this we view nullable value types as a great example of how to imbue a language-embraces special type with more permissive conversion semantics.
As a guiding principle, we would like tuple conversions to apply whenever a tuple can be deconstructed and reassembled into the new tuple type:
``` c#
(string, byte) t1 = ...;
(object, int) t2 = t1; // Allowed, because the following is:
(var a, var b) = t1; // Deconstruct, and ...
(object, int) t2 = (a, b); // reassemble
```
One problem is that nullable value type conversions are rather complex. They affect many parts of the language. It'd be great if we could make tuple conversions simpler. There are two principles we can try to follow:
1. A tuple conversion is a specific _kind_ of conversion, and it allows specific _kinds_ of conversions on the elements
2. A tuple conversion works in a given setting if all of its element conversions would work in that setting
The latter is more general, more complex and possibly ultimately necessary. However, somewhat to our surprise, we found a definition along the former principle that we cannot immediately poke a hole in:
> An _implicit tuple conversion_ is a standard conversion. It applies between two tuple types of equal arity when there is _any_ implicit conversion between each corresponding pair of types.
(Similarly for explicit conversions).
The interesting part here is that it's a standard conversion, so it is able to be composed with user defined conversions. Yet, its elements are allowed to perform their own user defined conversions! It feels like something could go wrong here, with recursive or circular application of user defined conversions, but we haven't been able to pinpoint an example.
A definition like this would be very desirable, because it won't require so much special casing around the spec.
We will try to implement this and see if we run into problems.
# Tuple-like construction of non-tuple types
We previously discussed to what extent non-tuple types should benefit from the tuple syntax. We've already decided that the deconstruction syntax applies to any type with a deconstructor, not just tuples. So what about construction?
The problem with allowing tuple literal syntax to construct any type is that _all_ types have constructors! There's no opt-in. This seems too out of control. Furthermore, it doesn't look intuitive that any old type can be "constructed" with a tuple literal:
``` c#
Dictionary<int, string> d = (16, EqualityComparer<int>.Default); / Huh???
```
This only seems meaningful if the constructor arguments coming in through a "tuple literal" are actually the constituent data of the object being created.
Finally, we don't have syntax for 0 and 1-tuples, so unless we add that, this would only even work when there's more than one constructor argument to the target type.
All in all, we don't think tuple literals should work for any types other than the built-in tuples. Instead, we want to brush off a feature that we've looked at before; the ability to omit the type from an object creation expression, when there is a target type:
``` c#
Point p = new (3, 4); // Same as new Point(3, 4)
List<string> l1 = new (10); // Works for 0 or 1 argument
List<int> l2 = new (){ 3, 4, 5 }; // Works with object/collection initializers, but must have parens as well.
```
Syntactically we would say that an object creation expression can omit the type when it has a parenthesized argument list. In the case of object and collection initializers, you cannot omit both the type and the parenthesized argument list, since that would lead to ambiguity with anonymous objects.
We think that this is promising. It is generally useful, and it would work nicely in the case of existing tuple-like types such as `System.Tuple<...>` and `KeyValuePair<...>`.
|
2.0
|
C# Design Notes for May 3-4, 2016 - # C# Design Notes for May 3-4, 2016
This pair of meetings further explored the space around tuple syntax, pattern matching and deconstruction.
1. Deconstructors - how to specify them
2. Switch conversions - how to deal with them
3. Tuple conversions - how to do them
4. Tuple-like types - how to construct them
Lots of concrete decisions, that allow us to make progress on implementation.
# Deconstructors
In #11031 we discussed the different contexts in which deconstruction should be able to occur, namely deconstructing _assignment_ (into existing variables), _declaration_ (into freshly declared local variables) and _patterns_ (as part of applying a recursive pattern).
We also explored the design space of how exactly "deconstructability" should be specified for a given type, but left the decision open - until now. Here's what we decided - and why. We'll stick to these decisions in initial prototypes, but as always are willing to be swayed by evidence as we roll them out and get usage.
**_Deconstruction should be specified with an instance (or extension) method**_. This is in keeping with other API patterns added throughout the history of C#, such as `GetEnumerator`, `Add`, and `GetAwaiter`. The benefit is that this leads to a relatively natural kind of member to have, and it can be specified with an extension method so that existing types can be augmented to be deconstructable outside of their own code.
The choice limits the ability of the pattern to later grow up to facilitate "active patterns". We aren't too concerned about that, because if we want to add active patterns at a later date we can easily come up with a separate mechanism for specifying those.
**_The instance/extension method should be called `Deconstruct`**_. We've been informally calling it `GetValues` for a while, but that name suffers from being in too popular use already, and not always for a similar purpose. This is a decision we're willing to alter if a better name comes along, and is sufficiently unencumbered.
**_The Deconstruct method "returns" the component values by use of individual out parameters**_. This choice may seem odd: after all we're adding a perfectly great feature called tuples, just so that you can return multiple values! The motivation here is primarily that we want `Deconstruct` to be overloadable. Sometimes there are genuinely multiple ways to deconstruct, and sometimes the type evolves over time to add more properties, and as you extend the `Deconstruct` method you also want to leave an old overload available for source and binary compat.
This one does nag us a little, because the declaration form with tuples is so much simpler, and would be sufficient in a majority of cases. On the other hand, this allows us to declare decomposition logic _for_ tuples the same way as for other types, which we couldn't if we depended on tuples for it!
Should this become a major nuisance (we don't think so) one could consider a hybrid approach where both tuple-returning and out-parameter versions were recognized, but for now we won't.
All in all, the deconstructor pattern looks like one of these:
``` c#
class Name
{
public void Deconstruct(out string first, out string last) { first = First; last = Last; }
...
}
// or
static class Extensions
{
public static void Deconstruct(this Name name, out string first, out string last) { first = name.First; last = name.Last; }
}
```
# Switch conversions
Switch statements today have a wrinkle where they will apply a unique implicit conversion from the switched-on expression to a (currently) switchable type. As we expand to allow switching on any type, this may be confusing at times, but we need to keep it at least in some scenarios, for backwards compatibility.
``` c#
switch (expr) // of some type Expression, which "cleverly" has a user defined conversion to int for evaluation
{
case Constant(int i): ... // Won't work, though Constant derives from Expression, because expr has been converted to int
...
}
```
Our current stance is that this is fringe enough for us to ignore. If you run into such a conversion and didn't want it, you'll have to work around it, e.g. by casting your switch expression to object.
If this turns out to be more of a nuisance we may have to come up with a smarter rule, but for now we're good with this.
# Tuple conversions
In #11031 we decided to add tuple conversions, that essentially convert tuples whenever their elements convert - unlike the more restrictive conversions that follow from `ValueTuple<...>` being a generic struct. In this we view nullable value types as a great example of how to imbue a language-embraces special type with more permissive conversion semantics.
As a guiding principle, we would like tuple conversions to apply whenever a tuple can be deconstructed and reassembled into the new tuple type:
``` c#
(string, byte) t1 = ...;
(object, int) t2 = t1; // Allowed, because the following is:
(var a, var b) = t1; // Deconstruct, and ...
(object, int) t2 = (a, b); // reassemble
```
One problem is that nullable value type conversions are rather complex. They affect many parts of the language. It'd be great if we could make tuple conversions simpler. There are two principles we can try to follow:
1. A tuple conversion is a specific _kind_ of conversion, and it allows specific _kinds_ of conversions on the elements
2. A tuple conversion works in a given setting if all of its element conversions would work in that setting
The latter is more general, more complex and possibly ultimately necessary. However, somewhat to our surprise, we found a definition along the former principle that we cannot immediately poke a hole in:
> An _implicit tuple conversion_ is a standard conversion. It applies between two tuple types of equal arity when there is _any_ implicit conversion between each corresponding pair of types.
(Similarly for explicit conversions).
The interesting part here is that it's a standard conversion, so it is able to be composed with user defined conversions. Yet, its elements are allowed to perform their own user defined conversions! It feels like something could go wrong here, with recursive or circular application of user defined conversions, but we haven't been able to pinpoint an example.
A definition like this would be very desirable, because it won't require so much special casing around the spec.
We will try to implement this and see if we run into problems.
# Tuple-like construction of non-tuple types
We previously discussed to what extent non-tuple types should benefit from the tuple syntax. We've already decided that the deconstruction syntax applies to any type with a deconstructor, not just tuples. So what about construction?
The problem with allowing tuple literal syntax to construct any type is that _all_ types have constructors! There's no opt-in. This seems too out of control. Furthermore, it doesn't look intuitive that any old type can be "constructed" with a tuple literal:
``` c#
Dictionary<int, string> d = (16, EqualityComparer<int>.Default); / Huh???
```
This only seems meaningful if the constructor arguments coming in through a "tuple literal" are actually the constituent data of the object being created.
Finally, we don't have syntax for 0 and 1-tuples, so unless we add that, this would only even work when there's more than one constructor argument to the target type.
All in all, we don't think tuple literals should work for any types other than the built-in tuples. Instead, we want to brush off a feature that we've looked at before; the ability to omit the type from an object creation expression, when there is a target type:
``` c#
Point p = new (3, 4); // Same as new Point(3, 4)
List<string> l1 = new (10); // Works for 0 or 1 argument
List<int> l2 = new (){ 3, 4, 5 }; // Works with object/collection initializers, but must have parens as well.
```
Syntactically we would say that an object creation expression can omit the type when it has a parenthesized argument list. In the case of object and collection initializers, you cannot omit both the type and the parenthesized argument list, since that would lead to ambiguity with anonymous objects.
We think that this is promising. It is generally useful, and it would work nicely in the case of existing tuple-like types such as `System.Tuple<...>` and `KeyValuePair<...>`.
|
non_test
|
c design notes for may c design notes for may this pair of meetings further explored the space around tuple syntax pattern matching and deconstruction deconstructors how to specify them switch conversions how to deal with them tuple conversions how to do them tuple like types how to construct them lots of concrete decisions that allow us to make progress on implementation deconstructors in we discussed the different contexts in which deconstruction should be able to occur namely deconstructing assignment into existing variables declaration into freshly declared local variables and patterns as part of applying a recursive pattern we also explored the design space of how exactly deconstructability should be specified for a given type but left the decision open until now here s what we decided and why we ll stick to these decisions in initial prototypes but as always are willing to be swayed by evidence as we roll them out and get usage deconstruction should be specified with an instance or extension method this is in keeping with other api patterns added throughout the history of c such as getenumerator add and getawaiter the benefit is that this leads to a relatively natural kind of member to have and it can be specified with an extension method so that existing types can be augmented to be deconstructable outside of their own code the choice limits the ability of the pattern to later grow up to facilitate active patterns we aren t too concerned about that because if we want to add active patterns at a later date we can easily come up with a separate mechanism for specifying those the instance extension method should be called deconstruct we ve been informally calling it getvalues for a while but that name suffers from being in too popular use already and not always for a similar purpose this is a decision we re willing to alter if a better name comes along and is sufficiently unencumbered the deconstruct method returns the component values by use of individual out parameters this choice may seem odd after all we re adding a perfectly great feature called tuples just so that you can return multiple values the motivation here is primarily that we want deconstruct to be overloadable sometimes there are genuinely multiple ways to deconstruct and sometimes the type evolves over time to add more properties and as you extend the deconstruct method you also want to leave an old overload available for source and binary compat this one does nag us a little because the declaration form with tuples is so much simpler and would be sufficient in a majority of cases on the other hand this allows us to declare decomposition logic for tuples the same way as for other types which we couldn t if we depended on tuples for it should this become a major nuisance we don t think so one could consider a hybrid approach where both tuple returning and out parameter versions were recognized but for now we won t all in all the deconstructor pattern looks like one of these c class name public void deconstruct out string first out string last first first last last or static class extensions public static void deconstruct this name name out string first out string last first name first last name last switch conversions switch statements today have a wrinkle where they will apply a unique implicit conversion from the switched on expression to a currently switchable type as we expand to allow switching on any type this may be confusing at times but we need to keep it at least in some scenarios for backwards compatibility c switch expr of some type expression which cleverly has a user defined conversion to int for evaluation case constant int i won t work though constant derives from expression because expr has been converted to int our current stance is that this is fringe enough for us to ignore if you run into such a conversion and didn t want it you ll have to work around it e g by casting your switch expression to object if this turns out to be more of a nuisance we may have to come up with a smarter rule but for now we re good with this tuple conversions in we decided to add tuple conversions that essentially convert tuples whenever their elements convert unlike the more restrictive conversions that follow from valuetuple being a generic struct in this we view nullable value types as a great example of how to imbue a language embraces special type with more permissive conversion semantics as a guiding principle we would like tuple conversions to apply whenever a tuple can be deconstructed and reassembled into the new tuple type c string byte object int allowed because the following is var a var b deconstruct and object int a b reassemble one problem is that nullable value type conversions are rather complex they affect many parts of the language it d be great if we could make tuple conversions simpler there are two principles we can try to follow a tuple conversion is a specific kind of conversion and it allows specific kinds of conversions on the elements a tuple conversion works in a given setting if all of its element conversions would work in that setting the latter is more general more complex and possibly ultimately necessary however somewhat to our surprise we found a definition along the former principle that we cannot immediately poke a hole in an implicit tuple conversion is a standard conversion it applies between two tuple types of equal arity when there is any implicit conversion between each corresponding pair of types similarly for explicit conversions the interesting part here is that it s a standard conversion so it is able to be composed with user defined conversions yet its elements are allowed to perform their own user defined conversions it feels like something could go wrong here with recursive or circular application of user defined conversions but we haven t been able to pinpoint an example a definition like this would be very desirable because it won t require so much special casing around the spec we will try to implement this and see if we run into problems tuple like construction of non tuple types we previously discussed to what extent non tuple types should benefit from the tuple syntax we ve already decided that the deconstruction syntax applies to any type with a deconstructor not just tuples so what about construction the problem with allowing tuple literal syntax to construct any type is that all types have constructors there s no opt in this seems too out of control furthermore it doesn t look intuitive that any old type can be constructed with a tuple literal c dictionary d equalitycomparer default huh this only seems meaningful if the constructor arguments coming in through a tuple literal are actually the constituent data of the object being created finally we don t have syntax for and tuples so unless we add that this would only even work when there s more than one constructor argument to the target type all in all we don t think tuple literals should work for any types other than the built in tuples instead we want to brush off a feature that we ve looked at before the ability to omit the type from an object creation expression when there is a target type c point p new same as new point list new works for or argument list new works with object collection initializers but must have parens as well syntactically we would say that an object creation expression can omit the type when it has a parenthesized argument list in the case of object and collection initializers you cannot omit both the type and the parenthesized argument list since that would lead to ambiguity with anonymous objects we think that this is promising it is generally useful and it would work nicely in the case of existing tuple like types such as system tuple and keyvaluepair
| 0
|
6,207
| 5,314,062,584
|
IssuesEvent
|
2017-02-13 14:07:29
|
woocommerce/woocommerce
|
https://api.github.com/repos/woocommerce/woocommerce
|
closed
|
Superfluous index on wp_woocommerce_tax_rate_locations table
|
Performance
|
## EXPLANATION OF THE ISSUE
The current definition of the woocommerce tax rate locations table (in the 2.7 master branch) is:
CREATE TABLE {$wpdb->prefix}woocommerce_tax_rate_locations (
location_id bigint(20) NOT NULL auto_increment,
location_code varchar(255) NOT NULL,
tax_rate_id bigint(20) NOT NULL,
location_type varchar(40) NOT NULL,
PRIMARY KEY (location_id),
KEY tax_rate_id (tax_rate_id),
KEY location_type (location_type),
KEY location_type_code (location_type(40),location_code(90))
) $collate;
We see that location_type is the first column in two indexes. Admittedly the second instance is technically a prefix index (only the first 40 characters are indexed), but the actual column definition is also VARCHAR(40) so actually that 40 character prefix is the entire column length.
Either way, the location_type index is superfluous. It can be removed without any impact other than saving some disk space.
Note that even if the column definition was longer, the 40 long prefix would be quite selective already and thus work as well as the separate index. The server would just use either index arbitrarily, and always pick the location_type_code indexes if the location_code column were specified in the WHERE clause.
Naturally, the story would be different if location_type were UNIQUE. Then it would serve a purpose. But that's not what we want here.
|
True
|
Superfluous index on wp_woocommerce_tax_rate_locations table - ## EXPLANATION OF THE ISSUE
The current definition of the woocommerce tax rate locations table (in the 2.7 master branch) is:
CREATE TABLE {$wpdb->prefix}woocommerce_tax_rate_locations (
location_id bigint(20) NOT NULL auto_increment,
location_code varchar(255) NOT NULL,
tax_rate_id bigint(20) NOT NULL,
location_type varchar(40) NOT NULL,
PRIMARY KEY (location_id),
KEY tax_rate_id (tax_rate_id),
KEY location_type (location_type),
KEY location_type_code (location_type(40),location_code(90))
) $collate;
We see that location_type is the first column in two indexes. Admittedly the second instance is technically a prefix index (only the first 40 characters are indexed), but the actual column definition is also VARCHAR(40) so actually that 40 character prefix is the entire column length.
Either way, the location_type index is superfluous. It can be removed without any impact other than saving some disk space.
Note that even if the column definition was longer, the 40 long prefix would be quite selective already and thus work as well as the separate index. The server would just use either index arbitrarily, and always pick the location_type_code indexes if the location_code column were specified in the WHERE clause.
Naturally, the story would be different if location_type were UNIQUE. Then it would serve a purpose. But that's not what we want here.
|
non_test
|
superfluous index on wp woocommerce tax rate locations table explanation of the issue the current definition of the woocommerce tax rate locations table in the master branch is create table wpdb prefix woocommerce tax rate locations location id bigint not null auto increment location code varchar not null tax rate id bigint not null location type varchar not null primary key location id key tax rate id tax rate id key location type location type key location type code location type location code collate we see that location type is the first column in two indexes admittedly the second instance is technically a prefix index only the first characters are indexed but the actual column definition is also varchar so actually that character prefix is the entire column length either way the location type index is superfluous it can be removed without any impact other than saving some disk space note that even if the column definition was longer the long prefix would be quite selective already and thus work as well as the separate index the server would just use either index arbitrarily and always pick the location type code indexes if the location code column were specified in the where clause naturally the story would be different if location type were unique then it would serve a purpose but that s not what we want here
| 0
|
9,825
| 3,073,506,944
|
IssuesEvent
|
2015-08-19 22:22:26
|
fsprojects/YC.PrettyPrinter
|
https://api.github.com/repos/fsprojects/YC.PrettyPrinter
|
closed
|
Fix links in documentation
|
SimpleTestTask task
|
Replace reference to fsharp.org with reference to http://yaccconstructor.github.io/ in gh-pages documentation.
|
1.0
|
Fix links in documentation - Replace reference to fsharp.org with reference to http://yaccconstructor.github.io/ in gh-pages documentation.
|
test
|
fix links in documentation replace reference to fsharp org with reference to in gh pages documentation
| 1
|
287,746
| 8,819,941,601
|
IssuesEvent
|
2019-01-01 04:02:39
|
Disalg-ICS-NJU/tlaplus-lamport-projects
|
https://api.github.com/repos/Disalg-ICS-NJU/tlaplus-lamport-projects
|
closed
|
JupiterInterface: How to Modify `state` in the Interface Level?
|
Rearrangement model check priority:normal refactor todo
|
***JupiterInterface: How to Modify `state` in the Interface Level?*** (Git Branch: Jupiter-History)
**Note: The following refactor makes JupiterInterface executable. How to take advantage of it?**
- [x] In `JupiterInterface`: Using `aop` to modify `state`
- [x] + New variable `aop[r]`: the actual operation applied at replica r
- [x] `state[r]` changed by applying `op[r]`: + `ApplyNewAop(r)`
- [x] `DoInt(c)`
- [x] `RevInt(c)`
- [x] `SRevInt`
- [x] In Jupiter protocols: Setting `op'` instead of modifying `state` directly
- [x] `AbsJupiter`
- [x] `CJupiter`
- [x] `CJupiterImplAbsJupiter`
- [x] `XJupiter`
- [x] `XJupiterExtended`
- [x] `XJupiterImplCJupiter`
- [x] `AJupiter`
- [x] `AJupiterExtended`
- [x] `AJupiterImplXJupiter`
Related Issues:
- #45 (Jupiter: Refactor `Do(c)`)
- #24 (Jupiter: To Separate Interface)
|
1.0
|
JupiterInterface: How to Modify `state` in the Interface Level? - ***JupiterInterface: How to Modify `state` in the Interface Level?*** (Git Branch: Jupiter-History)
**Note: The following refactor makes JupiterInterface executable. How to take advantage of it?**
- [x] In `JupiterInterface`: Using `aop` to modify `state`
- [x] + New variable `aop[r]`: the actual operation applied at replica r
- [x] `state[r]` changed by applying `op[r]`: + `ApplyNewAop(r)`
- [x] `DoInt(c)`
- [x] `RevInt(c)`
- [x] `SRevInt`
- [x] In Jupiter protocols: Setting `op'` instead of modifying `state` directly
- [x] `AbsJupiter`
- [x] `CJupiter`
- [x] `CJupiterImplAbsJupiter`
- [x] `XJupiter`
- [x] `XJupiterExtended`
- [x] `XJupiterImplCJupiter`
- [x] `AJupiter`
- [x] `AJupiterExtended`
- [x] `AJupiterImplXJupiter`
Related Issues:
- #45 (Jupiter: Refactor `Do(c)`)
- #24 (Jupiter: To Separate Interface)
|
non_test
|
jupiterinterface how to modify state in the interface level jupiterinterface how to modify state in the interface level git branch jupiter history note the following refactor makes jupiterinterface executable how to take advantage of it in jupiterinterface using aop to modify state new variable aop the actual operation applied at replica r state changed by applying op applynewaop r doint c revint c srevint in jupiter protocols setting op instead of modifying state directly absjupiter cjupiter cjupiterimplabsjupiter xjupiter xjupiterextended xjupiterimplcjupiter ajupiter ajupiterextended ajupiterimplxjupiter related issues jupiter refactor do c jupiter to separate interface
| 0
|
75,718
| 3,471,442,931
|
IssuesEvent
|
2015-12-23 15:25:07
|
wordpress-mobile/WordPress-Android
|
https://api.github.com/repos/wordpress-mobile/WordPress-Android
|
closed
|
Add a utility method to check if a URL is wpcom
|
core priority-high [Type] Task
|
While reviewing https://github.com/wordpress-mobile/WordPress-Android/pull/2784, I noticed we're using different way to check for a `wordpress.com` url across the app (hostname ends with `wordpress.com`, url contains `wordpress.com`).
|
1.0
|
Add a utility method to check if a URL is wpcom - While reviewing https://github.com/wordpress-mobile/WordPress-Android/pull/2784, I noticed we're using different way to check for a `wordpress.com` url across the app (hostname ends with `wordpress.com`, url contains `wordpress.com`).
|
non_test
|
add a utility method to check if a url is wpcom while reviewing i noticed we re using different way to check for a wordpress com url across the app hostname ends with wordpress com url contains wordpress com
| 0
|
99,095
| 4,046,924,912
|
IssuesEvent
|
2016-05-23 00:12:03
|
seiyria/deck.zone
|
https://api.github.com/repos/seiyria/deck.zone
|
closed
|
onbeforeunload warning
|
feature:standard priority:anytime
|
flag the editor as dirty before save and not dirty after a successful save
if the editor is dirty, closing the page should result in a warning
|
1.0
|
onbeforeunload warning - flag the editor as dirty before save and not dirty after a successful save
if the editor is dirty, closing the page should result in a warning
|
non_test
|
onbeforeunload warning flag the editor as dirty before save and not dirty after a successful save if the editor is dirty closing the page should result in a warning
| 0
|
216,072
| 24,215,267,161
|
IssuesEvent
|
2022-09-26 06:00:25
|
nidhi7598/linux-4.19.72_CVE-2022-29581
|
https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-29581
|
opened
|
CVE-2022-21499 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2022-21499 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.259</b>, <b>linuxlinux-4.19.259</b>, <b>linuxlinux-4.19.259</b>, <b>linuxlinux-4.19.259</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
KGDB and KDB allow read and write access to kernel memory, and thus should be restricted during lockdown. An attacker with access to a serial port could trigger the debugger so it is important that the debugger respect the lockdown mode when/if it is triggered. CVSS 3.1 Base Score 6.5 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.1/AV:L/AC:L/PR:H/UI:R/S:U/C:H/I:H/A:H).
<p>Publish Date: 2022-06-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21499>CVE-2022-21499</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-21499">https://www.linuxkernelcves.com/cves/CVE-2022-21499</a></p>
<p>Release Date: 2022-06-09</p>
<p>Fix Resolution: v5.4.197,v5.10.119,v5.15.42,v5.17.10,v5.18.1,v5.19-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-21499 (Medium) detected in multiple libraries - ## CVE-2022-21499 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.259</b>, <b>linuxlinux-4.19.259</b>, <b>linuxlinux-4.19.259</b>, <b>linuxlinux-4.19.259</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
KGDB and KDB allow read and write access to kernel memory, and thus should be restricted during lockdown. An attacker with access to a serial port could trigger the debugger so it is important that the debugger respect the lockdown mode when/if it is triggered. CVSS 3.1 Base Score 6.5 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.1/AV:L/AC:L/PR:H/UI:R/S:U/C:H/I:H/A:H).
<p>Publish Date: 2022-06-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21499>CVE-2022-21499</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-21499">https://www.linuxkernelcves.com/cves/CVE-2022-21499</a></p>
<p>Release Date: 2022-06-09</p>
<p>Fix Resolution: v5.4.197,v5.10.119,v5.15.42,v5.17.10,v5.18.1,v5.19-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries linuxlinux linuxlinux linuxlinux linuxlinux vulnerability details kgdb and kdb allow read and write access to kernel memory and thus should be restricted during lockdown an attacker with access to a serial port could trigger the debugger so it is important that the debugger respect the lockdown mode when if it is triggered cvss base score confidentiality integrity and availability impacts cvss vector cvss av l ac l pr h ui r s u c h i h a h publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
511,391
| 14,859,477,422
|
IssuesEvent
|
2021-01-18 18:33:44
|
InfiniteFlightAirportEditing/Airports
|
https://api.github.com/repos/InfiniteFlightAirportEditing/Airports
|
closed
|
FCBB-Maya Maya Airport-BRAZZAVILLE-REPUBLIC OF THE CONGO
|
Being Redone Priority 2
|
# Airport Name
Maya Maya Airport
# Country?
Republic Of The Congo
# Improvements that need to be made?
Redone
# Are you working on this airport?
Yes
# Airport Priority? (IF Event, 10000ft+ Runway, World/US Capital, Low)
World Capital
|
1.0
|
FCBB-Maya Maya Airport-BRAZZAVILLE-REPUBLIC OF THE CONGO - # Airport Name
Maya Maya Airport
# Country?
Republic Of The Congo
# Improvements that need to be made?
Redone
# Are you working on this airport?
Yes
# Airport Priority? (IF Event, 10000ft+ Runway, World/US Capital, Low)
World Capital
|
non_test
|
fcbb maya maya airport brazzaville republic of the congo airport name maya maya airport country republic of the congo improvements that need to be made redone are you working on this airport yes airport priority if event runway world us capital low world capital
| 0
|
140,646
| 11,354,308,429
|
IssuesEvent
|
2020-01-24 17:18:52
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
After excluding a site from A-C table `Show All` link is not shown in ac table
|
OS/Windows QA/Test-Plan-Specified QA/Yes bug feature/rewards regression
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
After excluding a site from A-C table `Show All` link is not shown
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.3.97
2. Enable rewards
3. Add a site to a-c table
4. Exclude the added site by clicking on Trash icon
5. After excluding a site from A-C table `Show All` link is not shown
## Actual result:
<!--Please add screenshots if needed-->
After excluding a site from A-C table `Show All` link is not shown in ac table

## Expected result:
After excluding a site from A-C table `Show All` link should be shown in a-c table
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Always
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.3.97 Chromium: 79.0.3945.130 (Official Build) beta (64-bit)
-- | --
Revision | e22de67c28798d98833a7137c0e22876237fc40a-refs/branch-heads/3945@{#1047}
OS | Windows 10 OS Version 1803 (Build 17134.1006)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? No
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the dev channel? No
- Can you reproduce this issue with the nightly channel? No
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @brave/legacy_qa @NejcZdovc @rebron
|
1.0
|
After excluding a site from A-C table `Show All` link is not shown in ac table - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
After excluding a site from A-C table `Show All` link is not shown
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.3.97
2. Enable rewards
3. Add a site to a-c table
4. Exclude the added site by clicking on Trash icon
5. After excluding a site from A-C table `Show All` link is not shown
## Actual result:
<!--Please add screenshots if needed-->
After excluding a site from A-C table `Show All` link is not shown in ac table

## Expected result:
After excluding a site from A-C table `Show All` link should be shown in a-c table
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Always
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.3.97 Chromium: 79.0.3945.130 (Official Build) beta (64-bit)
-- | --
Revision | e22de67c28798d98833a7137c0e22876237fc40a-refs/branch-heads/3945@{#1047}
OS | Windows 10 OS Version 1803 (Build 17134.1006)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? No
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the dev channel? No
- Can you reproduce this issue with the nightly channel? No
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @brave/legacy_qa @NejcZdovc @rebron
|
test
|
after excluding a site from a c table show all link is not shown in ac table have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description after excluding a site from a c table show all link is not shown steps to reproduce clean profile enable rewards add a site to a c table exclude the added site by clicking on trash icon after excluding a site from a c table show all link is not shown actual result after excluding a site from a c table show all link is not shown in ac table expected result after excluding a site from a c table show all link should be shown in a c table reproduces how often always brave version brave version info brave chromium official build beta bit revision refs branch heads os windows os version build version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel yes can you reproduce this issue with the dev channel no can you reproduce this issue with the nightly channel no other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave legacy qa nejczdovc rebron
| 1
|
156,345
| 12,306,141,492
|
IssuesEvent
|
2020-05-12 00:31:39
|
cosinekitty/astronomy
|
https://api.github.com/repos/cosinekitty/astronomy
|
closed
|
Unit tests: make it more obvious which language is being tested
|
testing
|
When running the unit tests, it is not always obvious which of the 4 supported languages is currently being tested. It would be nice to make it more obvious. Ideas:
- Put a big ASCII art banner in front of each language test.
- Consistently prefix all test output with the language name.
|
1.0
|
Unit tests: make it more obvious which language is being tested - When running the unit tests, it is not always obvious which of the 4 supported languages is currently being tested. It would be nice to make it more obvious. Ideas:
- Put a big ASCII art banner in front of each language test.
- Consistently prefix all test output with the language name.
|
test
|
unit tests make it more obvious which language is being tested when running the unit tests it is not always obvious which of the supported languages is currently being tested it would be nice to make it more obvious ideas put a big ascii art banner in front of each language test consistently prefix all test output with the language name
| 1
|
248,580
| 21,042,032,320
|
IssuesEvent
|
2022-03-31 13:10:01
|
nrwl/nx
|
https://api.github.com/repos/nrwl/nx
|
closed
|
Don't error on "targetDependencies" "self" in nx.json
|
type: feature blocked: retry with latest scope: core
|
<!-- Please do your best to fill out all of the sections below! -->
<!-- Use this issue type for concrete suggestions, otherwise, open a discussion type issue instead. -->
- [] I'd be willing to implement this feature ([contributing guide](https://github.com/nrwl/nx/blob/master/CONTRIBUTING.md))
## Description
<!-- What is the behavior that you would like to see introduced? -->
Currently, we have libs that lint and test depend on the codegen target that lives in those projects.
We have to make sure to add "dependsOn" in each of these libs to make sure "codegen" is run before lint and test.
We would like to instead be able to do this in `nx.json`, but currently nx throws an error if "targetDependencies" is "self" and the project doesn't contain that target.
## Motivation
<!-- Why do you believe this behavior would be beneficial? -->
Easier maintainability for depends on targets for "self".
## Suggested Implementation
<!-- How do you imagine this might work? -->
Don't throw an error if the "targetDependencies" "self" is not a target in the project.
```
...
"targetDependencies": {
"build": [
{
"target": "build",
"projects": "dependencies"
},
{
"target": "codegen",
"projects": "dependencies"
},
{
"target": "codegen",
"projects": "self" // < THROWS IF NOT TARGET IN CURRENT PROJECT
}
],
...
```
## Alternate Implementations
<!-- How else do you imagine this might work? -->
A new target like "self-optional" or something to indicate to not throw if the target is not there.
|
1.0
|
Don't error on "targetDependencies" "self" in nx.json - <!-- Please do your best to fill out all of the sections below! -->
<!-- Use this issue type for concrete suggestions, otherwise, open a discussion type issue instead. -->
- [] I'd be willing to implement this feature ([contributing guide](https://github.com/nrwl/nx/blob/master/CONTRIBUTING.md))
## Description
<!-- What is the behavior that you would like to see introduced? -->
Currently, we have libs that lint and test depend on the codegen target that lives in those projects.
We have to make sure to add "dependsOn" in each of these libs to make sure "codegen" is run before lint and test.
We would like to instead be able to do this in `nx.json`, but currently nx throws an error if "targetDependencies" is "self" and the project doesn't contain that target.
## Motivation
<!-- Why do you believe this behavior would be beneficial? -->
Easier maintainability for depends on targets for "self".
## Suggested Implementation
<!-- How do you imagine this might work? -->
Don't throw an error if the "targetDependencies" "self" is not a target in the project.
```
...
"targetDependencies": {
"build": [
{
"target": "build",
"projects": "dependencies"
},
{
"target": "codegen",
"projects": "dependencies"
},
{
"target": "codegen",
"projects": "self" // < THROWS IF NOT TARGET IN CURRENT PROJECT
}
],
...
```
## Alternate Implementations
<!-- How else do you imagine this might work? -->
A new target like "self-optional" or something to indicate to not throw if the target is not there.
|
test
|
don t error on targetdependencies self in nx json i d be willing to implement this feature description currently we have libs that lint and test depend on the codegen target that lives in those projects we have to make sure to add dependson in each of these libs to make sure codegen is run before lint and test we would like to instead be able to do this in nx json but currently nx throws an error if targetdependencies is self and the project doesn t contain that target motivation easier maintainability for depends on targets for self suggested implementation don t throw an error if the targetdependencies self is not a target in the project targetdependencies build target build projects dependencies target codegen projects dependencies target codegen projects self throws if not target in current project alternate implementations a new target like self optional or something to indicate to not throw if the target is not there
| 1
|
43,910
| 5,716,435,082
|
IssuesEvent
|
2017-04-19 15:08:23
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
move towards a modal confirmation/alert dialog rather than native alerts/confirmations
|
:Design :Platform enhancement
|
this is mainly for discussion, but I would like to see Kibana move towards a more integrated alert/confirmation UI rather than depending on the native alert/confirm elements.
Native browser confirmations/alerts are almost becoming an anti-pattern; evidenced by the fact that lots of browsers start adding the "do you want to prevent more popups" option after one of two times.
|
1.0
|
move towards a modal confirmation/alert dialog rather than native alerts/confirmations - this is mainly for discussion, but I would like to see Kibana move towards a more integrated alert/confirmation UI rather than depending on the native alert/confirm elements.
Native browser confirmations/alerts are almost becoming an anti-pattern; evidenced by the fact that lots of browsers start adding the "do you want to prevent more popups" option after one of two times.
|
non_test
|
move towards a modal confirmation alert dialog rather than native alerts confirmations this is mainly for discussion but i would like to see kibana move towards a more integrated alert confirmation ui rather than depending on the native alert confirm elements native browser confirmations alerts are almost becoming an anti pattern evidenced by the fact that lots of browsers start adding the do you want to prevent more popups option after one of two times
| 0
|
127,557
| 10,474,955,255
|
IssuesEvent
|
2019-09-23 15:24:22
|
paritytech/parity-ethereum
|
https://api.github.com/repos/paritytech/parity-ethereum
|
closed
|
[json-tests] stStaticCall blockchain test failures
|
F4-tests 💻 P2-asap 🌊
|
After https://github.com/paritytech/parity-ethereum/pull/11054 the following JSON tests fail in [`stStaticCall`](https://github.com/ethereum/tests/blob/f386934ce509fa0a4dbd1c68421b3361b01b2efa/BlockchainTests/GeneralStateTests/stStaticCall/static_callBasic.json#L388) and are added to the list of skipped tests in `ethcore/res/ethereum/tests-issues/currents.json`.
```
"static_callBasic_d0g0v0_ConstantinopleFix",
"static_callBasic_d1g0v0_ConstantinopleFix",
"static_callBasic_d2g0v0_ConstantinopleFix",
"static_callBasic_d3g0v0_ConstantinopleFix"
```
Fix the tests and "un-skip" them.
|
1.0
|
[json-tests] stStaticCall blockchain test failures - After https://github.com/paritytech/parity-ethereum/pull/11054 the following JSON tests fail in [`stStaticCall`](https://github.com/ethereum/tests/blob/f386934ce509fa0a4dbd1c68421b3361b01b2efa/BlockchainTests/GeneralStateTests/stStaticCall/static_callBasic.json#L388) and are added to the list of skipped tests in `ethcore/res/ethereum/tests-issues/currents.json`.
```
"static_callBasic_d0g0v0_ConstantinopleFix",
"static_callBasic_d1g0v0_ConstantinopleFix",
"static_callBasic_d2g0v0_ConstantinopleFix",
"static_callBasic_d3g0v0_ConstantinopleFix"
```
Fix the tests and "un-skip" them.
|
test
|
ststaticcall blockchain test failures after the following json tests fail in and are added to the list of skipped tests in ethcore res ethereum tests issues currents json static callbasic constantinoplefix static callbasic constantinoplefix static callbasic constantinoplefix static callbasic constantinoplefix fix the tests and un skip them
| 1
|
9,307
| 3,035,265,212
|
IssuesEvent
|
2015-08-06 01:25:52
|
hahndorf/Set-Privacy
|
https://api.github.com/repos/hahndorf/Set-Privacy
|
closed
|
Disable telemetry
|
enhancement help wanted needs testing
|
fix10.isleaked.com suggests:
sc delete DiagTrack
sc delete dmwappushservice
echo "" > C:\ProgramData\Microsoft\Diagnosis\ETLLogs\AutoLogger\AutoLogger-Diagtrack-Listener.etl
reg add "HKLM\SOFTWARE\Policies\Microsoft\Windows\DataCollection" /v AllowTelemetry /t REG_DWORD /d 0 /f
I would rather disable the services.
|
1.0
|
Disable telemetry - fix10.isleaked.com suggests:
sc delete DiagTrack
sc delete dmwappushservice
echo "" > C:\ProgramData\Microsoft\Diagnosis\ETLLogs\AutoLogger\AutoLogger-Diagtrack-Listener.etl
reg add "HKLM\SOFTWARE\Policies\Microsoft\Windows\DataCollection" /v AllowTelemetry /t REG_DWORD /d 0 /f
I would rather disable the services.
|
test
|
disable telemetry isleaked com suggests sc delete diagtrack sc delete dmwappushservice echo c programdata microsoft diagnosis etllogs autologger autologger diagtrack listener etl reg add hklm software policies microsoft windows datacollection v allowtelemetry t reg dword d f i would rather disable the services
| 1
|
194,377
| 14,676,651,007
|
IssuesEvent
|
2020-12-30 20:48:24
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
bigfish02/kruise: pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go; 3 LoC
|
fresh test tiny
|
Found a possible issue in [bigfish02/kruise](https://www.github.com/bigfish02/kruise) at [pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go](https://github.com/bigfish02/kruise/blob/7f68784556d41872df875eb5318b4b771728320e/pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go#L1622-L1624)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to ud at line 1623 may start a goroutine
[Click here to see the code in its original context.](https://github.com/bigfish02/kruise/blob/7f68784556d41872df875eb5318b4b771728320e/pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go#L1622-L1624)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ud := range udList.Items {
c.Delete(context.TODO(), &ud)
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
The following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine:
digraph G {
"(configureTransport, 1)" -> {"(addConnIfNeeded, 3)";}
"(createEncoderOfType, 2)" -> {"(_createEncoderOfType, 2)";}
"(Delete, 2)" -> {"(Add, 1)";}
"(encoderOfMap, 2)" -> {"(append, 1)";}
"(add, 2)" -> {"(findObject, 2)";}
"(get, 1)" -> {"(SetTransportDefaults, 1)";}
"(ConfigureTransport, 1)" -> {"(configureTransport, 1)";}
"(encoderOfType, 2)" -> {"(createEncoderOfType, 2)";}
"(append, 1)" -> {}
"(newClient, 1)" -> {}
"(findObject, 2)" -> {"(get, 1)";}
"(_createEncoderOfType, 2)" -> {"(encoderOfArray, 2)";"(encoderOfSlice, 2)";"(encoderOfMap, 2)";"(createEncoderOfNative, 2)";"(encoderOfStruct, 2)";}
"(encoderOfSlice, 2)" -> {"(append, 1)";}
"(New, 1)" -> {"(newClient, 1)";"(get, 1)";}
"(addConnIfNeeded, 3)" -> {}
"(Encode, 2)" -> {"(encoderOfMapKey, 2)";}
"(encoderOfArray, 2)" -> {"(append, 1)";}
"(encoderOfStruct, 2)" -> {"(describeStruct, 2)";}
"(SetTransportDefaults, 1)" -> {"(ConfigureTransport, 1)";}
"(Update, 2)" -> {"(Encode, 2)";}
"(encoderOfMapKey, 2)" -> {"(encoderOfType, 2)";}
"(Update, 1)" -> {"(Update, 2)";}
"(add, 1)" -> {"(Update, 1)";}
"(Push, 2)" -> {"(Update, 1)";}
"(createEncoderOfNative, 2)" -> {"(decoderOfSlice, 2)";}
"(decoderOfSlice, 2)" -> {"(append, 1)";}
"(describeStruct, 2)" -> {"(append, 1)";}
"(Add, 1)" -> {"(Push, 2)";"(New, 1)";"(add, 2)";"(add, 1)";}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 7f68784556d41872df875eb5318b4b771728320e
|
1.0
|
bigfish02/kruise: pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go; 3 LoC -
Found a possible issue in [bigfish02/kruise](https://www.github.com/bigfish02/kruise) at [pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go](https://github.com/bigfish02/kruise/blob/7f68784556d41872df875eb5318b4b771728320e/pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go#L1622-L1624)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to ud at line 1623 may start a goroutine
[Click here to see the code in its original context.](https://github.com/bigfish02/kruise/blob/7f68784556d41872df875eb5318b4b771728320e/pkg/controller/uniteddeployment/uniteddeployment_controller_statefulset_test.go#L1622-L1624)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ud := range udList.Items {
c.Delete(context.TODO(), &ud)
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
The following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine:
digraph G {
"(configureTransport, 1)" -> {"(addConnIfNeeded, 3)";}
"(createEncoderOfType, 2)" -> {"(_createEncoderOfType, 2)";}
"(Delete, 2)" -> {"(Add, 1)";}
"(encoderOfMap, 2)" -> {"(append, 1)";}
"(add, 2)" -> {"(findObject, 2)";}
"(get, 1)" -> {"(SetTransportDefaults, 1)";}
"(ConfigureTransport, 1)" -> {"(configureTransport, 1)";}
"(encoderOfType, 2)" -> {"(createEncoderOfType, 2)";}
"(append, 1)" -> {}
"(newClient, 1)" -> {}
"(findObject, 2)" -> {"(get, 1)";}
"(_createEncoderOfType, 2)" -> {"(encoderOfArray, 2)";"(encoderOfSlice, 2)";"(encoderOfMap, 2)";"(createEncoderOfNative, 2)";"(encoderOfStruct, 2)";}
"(encoderOfSlice, 2)" -> {"(append, 1)";}
"(New, 1)" -> {"(newClient, 1)";"(get, 1)";}
"(addConnIfNeeded, 3)" -> {}
"(Encode, 2)" -> {"(encoderOfMapKey, 2)";}
"(encoderOfArray, 2)" -> {"(append, 1)";}
"(encoderOfStruct, 2)" -> {"(describeStruct, 2)";}
"(SetTransportDefaults, 1)" -> {"(ConfigureTransport, 1)";}
"(Update, 2)" -> {"(Encode, 2)";}
"(encoderOfMapKey, 2)" -> {"(encoderOfType, 2)";}
"(Update, 1)" -> {"(Update, 2)";}
"(add, 1)" -> {"(Update, 1)";}
"(Push, 2)" -> {"(Update, 1)";}
"(createEncoderOfNative, 2)" -> {"(decoderOfSlice, 2)";}
"(decoderOfSlice, 2)" -> {"(append, 1)";}
"(describeStruct, 2)" -> {"(append, 1)";}
"(Add, 1)" -> {"(Push, 2)";"(New, 1)";"(add, 2)";"(add, 1)";}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 7f68784556d41872df875eb5318b4b771728320e
|
test
|
kruise pkg controller uniteddeployment uniteddeployment controller statefulset test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to ud at line may start a goroutine click here to show the line s of go which triggered the analyzer go for ud range udlist items c delete context todo ud click here to show extra information the analyzer produced the following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine digraph g configuretransport addconnifneeded createencoderoftype createencoderoftype delete add encoderofmap append add findobject get settransportdefaults configuretransport configuretransport encoderoftype createencoderoftype append newclient findobject get createencoderoftype encoderofarray encoderofslice encoderofmap createencoderofnative encoderofstruct encoderofslice append new newclient get addconnifneeded encode encoderofmapkey encoderofarray append encoderofstruct describestruct settransportdefaults configuretransport update encode encoderofmapkey encoderoftype update update add update push update createencoderofnative decoderofslice decoderofslice append describestruct append add push new add add leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
112,095
| 9,553,727,276
|
IssuesEvent
|
2019-05-02 20:03:07
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Flaky test: $scope.$watchMulti "before each" hook
|
test
|
```
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 605 of 725 SUCCESS (0 secs / 3.169 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) $scope.$watchMulti "before each" hook: workFn FAILED
13:32:27 Error: Uncaught TypeError: Cannot read property 'some' of undefined (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:99930)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 606 of 725 (1 FAILED) (0 secs / 3.189 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 607 of 725 (1 FAILED) (0 secs / 3.191 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) "after each" hook FAILED
13:32:27 TypeError: Cannot read property '$injector' of null
13:32:27 at Context.<anonymous> (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:101912:32)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 608 of 725 (2 FAILED) (0 secs / 3.192 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) "after each" hook FAILED
13:32:27 TypeError: Cannot read property '$injector' of null
13:32:27 at Context.<anonymous> (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:101912:32)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 609 of 725 (3 FAILED) (0 secs / 3.192 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) "after each" hook for "should show correct series" FAILED
13:32:27 TypeError: Cannot read property '$injector' of null
13:32:27 at Context.<anonymous> (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:101912:32)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 610 of 725 (4 FAILED) (0 secs / 3.192 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 610 of 725 (4 FAILED) (13.036 secs / 3.192 secs)
13:32:27 Warning: Task "karma:ciShard-3" failed. Use --force to continue.
```
|
1.0
|
Flaky test: $scope.$watchMulti "before each" hook - ```
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 605 of 725 SUCCESS (0 secs / 3.169 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) $scope.$watchMulti "before each" hook: workFn FAILED
13:32:27 Error: Uncaught TypeError: Cannot read property 'some' of undefined (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:99930)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 606 of 725 (1 FAILED) (0 secs / 3.189 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 607 of 725 (1 FAILED) (0 secs / 3.191 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) "after each" hook FAILED
13:32:27 TypeError: Cannot read property '$injector' of null
13:32:27 at Context.<anonymous> (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:101912:32)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 608 of 725 (2 FAILED) (0 secs / 3.192 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) "after each" hook FAILED
13:32:27 TypeError: Cannot read property '$injector' of null
13:32:27 at Context.<anonymous> (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:101912:32)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 609 of 725 (3 FAILED) (0 secs / 3.192 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0) "after each" hook for "should show correct series" FAILED
13:32:27 TypeError: Cannot read property '$injector' of null
13:32:27 at Context.<anonymous> (http://localhost:5610/bundles/tests.bundle.js?shards=4&shard_num=3:101912:32)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 610 of 725 (4 FAILED) (0 secs / 3.192 secs)
13:32:27 Chrome 59.0.3071 (Linux 0.0.0): Executed 610 of 725 (4 FAILED) (13.036 secs / 3.192 secs)
13:32:27 Warning: Task "karma:ciShard-3" failed. Use --force to continue.
```
|
test
|
flaky test scope watchmulti before each hook chrome linux executed of success secs secs chrome linux scope watchmulti before each hook workfn failed error uncaught typeerror cannot read property some of undefined chrome linux executed of failed secs secs chrome linux executed of failed secs secs chrome linux after each hook failed typeerror cannot read property injector of null at context chrome linux executed of failed secs secs chrome linux after each hook failed typeerror cannot read property injector of null at context chrome linux executed of failed secs secs chrome linux after each hook for should show correct series failed typeerror cannot read property injector of null at context chrome linux executed of failed secs secs chrome linux executed of failed secs secs warning task karma cishard failed use force to continue
| 1
|
8,922
| 3,010,900,883
|
IssuesEvent
|
2015-07-28 15:21:06
|
omeka/plugin-Commenting
|
https://api.github.com/repos/omeka/plugin-Commenting
|
closed
|
Confirm comments management
|
Testing
|
On admin side, go to Comments. Marks some spam and flag some others.
Should see gree checkmarks changing to red warning signs and back. Flagged comments should get red background.
Check the quick filtering.
Check that the bulk actions with checkboxes work.
Check deleting comments works. Also, delete an Item that has some comments and make sure its comments are deleted.
|
1.0
|
Confirm comments management - On admin side, go to Comments. Marks some spam and flag some others.
Should see gree checkmarks changing to red warning signs and back. Flagged comments should get red background.
Check the quick filtering.
Check that the bulk actions with checkboxes work.
Check deleting comments works. Also, delete an Item that has some comments and make sure its comments are deleted.
|
test
|
confirm comments management on admin side go to comments marks some spam and flag some others should see gree checkmarks changing to red warning signs and back flagged comments should get red background check the quick filtering check that the bulk actions with checkboxes work check deleting comments works also delete an item that has some comments and make sure its comments are deleted
| 1
|
22,276
| 18,943,530,871
|
IssuesEvent
|
2021-11-18 07:29:28
|
VirtusLab/git-machete
|
https://api.github.com/repos/VirtusLab/git-machete
|
closed
|
`github create-pr`: check if base branch for PR exists in remote
|
github usability
|
Attempt to create a pull request with base branch being already deleted from remote ends up with `Unprocessable Entity` error. (example in #332).
Proposed solution:
Perform `git fetch <remote>` at the beginning of `create-pr`. if base branch is not present in remote branches, perform `handle_untracked_branch` with relevant remote for missing base branch.
|
True
|
`github create-pr`: check if base branch for PR exists in remote - Attempt to create a pull request with base branch being already deleted from remote ends up with `Unprocessable Entity` error. (example in #332).
Proposed solution:
Perform `git fetch <remote>` at the beginning of `create-pr`. if base branch is not present in remote branches, perform `handle_untracked_branch` with relevant remote for missing base branch.
|
non_test
|
github create pr check if base branch for pr exists in remote attempt to create a pull request with base branch being already deleted from remote ends up with unprocessable entity error example in proposed solution perform git fetch at the beginning of create pr if base branch is not present in remote branches perform handle untracked branch with relevant remote for missing base branch
| 0
|
21,358
| 7,007,949,727
|
IssuesEvent
|
2017-12-19 14:13:35
|
furzeface/framewerk
|
https://api.github.com/repos/furzeface/framewerk
|
closed
|
Appveyor config
|
build chore feature
|
# Project version:
v1.0.0
## Issue type:
- [x] Feature suggestion
Add a sample AppVeyor config for deployment too along with Travis and Circle.
|
1.0
|
Appveyor config - # Project version:
v1.0.0
## Issue type:
- [x] Feature suggestion
Add a sample AppVeyor config for deployment too along with Travis and Circle.
|
non_test
|
appveyor config project version issue type feature suggestion add a sample appveyor config for deployment too along with travis and circle
| 0
|
146,874
| 11,759,748,624
|
IssuesEvent
|
2020-03-13 17:55:10
|
awslabs/s2n
|
https://api.github.com/repos/awslabs/s2n
|
opened
|
Integration tests leave s2nd and openssl processes behind
|
type/test
|
## **Problem:**
Integration tests run locally during development fail inconsistently with this pattern:
Running TLS1.3 handshake tests with openssl: OpenSSL 1.1.1e-dev xx XXX xxxx
Running scenarios: 24
Mode:server Version:TLS13 Curve:P-256 Cipher:TLS_AES_256_GCM_SHA384 PASSED
...
Mode:client Version:TLS13 Curve:P-256 Cipher:TLS_AES_256_GCM_SHA384 FAILED
openssl server: 139856980317248:error:02006062:system library:bind:Address already in use:crypto/bio/b_sock2.c:161:
139856980317248:error:20093075:BIO routines:BIO_bind:unable to bind socket:crypto/bio/b_sock2.c:162:
The current workaround is to use ps aux or ps -A to find whether extra s2nd or openssl processes are hanging around, and pkill -c [openssl|s2nd]
## **Proposed Solution:**
Manage ports used for testing, so processes that get stuck / deadlock can be cleaned up properly.
|
1.0
|
Integration tests leave s2nd and openssl processes behind - ## **Problem:**
Integration tests run locally during development fail inconsistently with this pattern:
Running TLS1.3 handshake tests with openssl: OpenSSL 1.1.1e-dev xx XXX xxxx
Running scenarios: 24
Mode:server Version:TLS13 Curve:P-256 Cipher:TLS_AES_256_GCM_SHA384 PASSED
...
Mode:client Version:TLS13 Curve:P-256 Cipher:TLS_AES_256_GCM_SHA384 FAILED
openssl server: 139856980317248:error:02006062:system library:bind:Address already in use:crypto/bio/b_sock2.c:161:
139856980317248:error:20093075:BIO routines:BIO_bind:unable to bind socket:crypto/bio/b_sock2.c:162:
The current workaround is to use ps aux or ps -A to find whether extra s2nd or openssl processes are hanging around, and pkill -c [openssl|s2nd]
## **Proposed Solution:**
Manage ports used for testing, so processes that get stuck / deadlock can be cleaned up properly.
|
test
|
integration tests leave and openssl processes behind problem integration tests run locally during development fail inconsistently with this pattern running handshake tests with openssl openssl dev xx xxx xxxx running scenarios mode server version curve p cipher tls aes gcm passed mode client version curve p cipher tls aes gcm failed openssl server error system library bind address already in use crypto bio b c error bio routines bio bind unable to bind socket crypto bio b c the current workaround is to use ps aux or ps a to find whether extra or openssl processes are hanging around and pkill c proposed solution manage ports used for testing so processes that get stuck deadlock can be cleaned up properly
| 1
|
144,847
| 5,546,786,052
|
IssuesEvent
|
2017-03-23 02:28:14
|
CS2103JAN2017-T11-B3/main
|
https://api.github.com/repos/CS2103JAN2017-T11-B3/main
|
closed
|
Update DeveloperGuide to reflect v0.5 objectives
|
priority.high
|
keep in mind from handout:
> Developer Guide: Update Developer Guide to match your current design or the design you plan to have at V0.5.
> [1] While it is OK for some sections to be written collectively, each student should have some sections that were written individually. For example, the section describing the design of Logic component should be written by the person in charge of the Logic component.
> [2] It is not enough to simply do minor tweaks to the existing document to match your project if your design is very similar to the sample given. We need to see a significant amount of content written by each student to judge how well you can document technical details and use relevant models.
> Here are some things you can add:
> Compare the current design with other alternative designs possible.
> Add more descriptions e.g. show at least one interaction within each component.
> Use object diagrams to show sample object structures at runtime (if it helps the reader).
> Explain how you implemented complicated features such as undo/redo.
> In future (after these topics have been covered):
> * Use Activity Diagrams and OO Domain Models where applicable.
> * Mention where design patterns/principles are applied.
|
1.0
|
Update DeveloperGuide to reflect v0.5 objectives - keep in mind from handout:
> Developer Guide: Update Developer Guide to match your current design or the design you plan to have at V0.5.
> [1] While it is OK for some sections to be written collectively, each student should have some sections that were written individually. For example, the section describing the design of Logic component should be written by the person in charge of the Logic component.
> [2] It is not enough to simply do minor tweaks to the existing document to match your project if your design is very similar to the sample given. We need to see a significant amount of content written by each student to judge how well you can document technical details and use relevant models.
> Here are some things you can add:
> Compare the current design with other alternative designs possible.
> Add more descriptions e.g. show at least one interaction within each component.
> Use object diagrams to show sample object structures at runtime (if it helps the reader).
> Explain how you implemented complicated features such as undo/redo.
> In future (after these topics have been covered):
> * Use Activity Diagrams and OO Domain Models where applicable.
> * Mention where design patterns/principles are applied.
|
non_test
|
update developerguide to reflect objectives keep in mind from handout developer guide update developer guide to match your current design or the design you plan to have at while it is ok for some sections to be written collectively each student should have some sections that were written individually for example the section describing the design of logic component should be written by the person in charge of the logic component it is not enough to simply do minor tweaks to the existing document to match your project if your design is very similar to the sample given we need to see a significant amount of content written by each student to judge how well you can document technical details and use relevant models here are some things you can add compare the current design with other alternative designs possible add more descriptions e g show at least one interaction within each component use object diagrams to show sample object structures at runtime if it helps the reader explain how you implemented complicated features such as undo redo in future after these topics have been covered use activity diagrams and oo domain models where applicable mention where design patterns principles are applied
| 0
|
144,178
| 11,596,967,104
|
IssuesEvent
|
2020-02-24 19:56:05
|
sertmer/GlobeTrotter
|
https://api.github.com/repos/sertmer/GlobeTrotter
|
closed
|
Test apiCalls
|
FrontEnd Sprint 2 Testing
|
Each fetch block requires the following:
- Test URL .toHaveBeenCalledWith()
- Returned value (happy path - resolves.toEqual)
- Returned value (sad path{s} - rejects.toEqual)
|
1.0
|
Test apiCalls - Each fetch block requires the following:
- Test URL .toHaveBeenCalledWith()
- Returned value (happy path - resolves.toEqual)
- Returned value (sad path{s} - rejects.toEqual)
|
test
|
test apicalls each fetch block requires the following test url tohavebeencalledwith returned value happy path resolves toequal returned value sad path s rejects toequal
| 1
|
493,101
| 14,226,214,948
|
IssuesEvent
|
2020-11-17 22:34:59
|
emre1702/TDS-V-Public
|
https://api.github.com/repos/emre1702/TDS-V-Public
|
closed
|
[BUG] Some gangs have no gang lobby teams
|
bug low priority
|
**Describe the bug**
Some gangs have no gang lobby teams. So trying to join the gang lobby crashes the server.
**Expected behavior**
Every gang needs exactly one gang lobby team.
**Screenshots / Logs**
at TDS_Server.LobbySystem.Players.GangLobbyPlayers.AddPlayer(ITDSPlayer player, Int32 teamIndex) in B:\Users\EmreKara\Desktop\Tools\GitHub\TDS-V\Server\Lobby\Players\GangLobbyPlayers.cs:line 19
Low priority because gang lobby isn't even opened yet.
|
1.0
|
[BUG] Some gangs have no gang lobby teams - **Describe the bug**
Some gangs have no gang lobby teams. So trying to join the gang lobby crashes the server.
**Expected behavior**
Every gang needs exactly one gang lobby team.
**Screenshots / Logs**
at TDS_Server.LobbySystem.Players.GangLobbyPlayers.AddPlayer(ITDSPlayer player, Int32 teamIndex) in B:\Users\EmreKara\Desktop\Tools\GitHub\TDS-V\Server\Lobby\Players\GangLobbyPlayers.cs:line 19
Low priority because gang lobby isn't even opened yet.
|
non_test
|
some gangs have no gang lobby teams describe the bug some gangs have no gang lobby teams so trying to join the gang lobby crashes the server expected behavior every gang needs exactly one gang lobby team screenshots logs at tds server lobbysystem players ganglobbyplayers addplayer itdsplayer player teamindex in b users emrekara desktop tools github tds v server lobby players ganglobbyplayers cs line low priority because gang lobby isn t even opened yet
| 0
|
61,564
| 6,743,005,022
|
IssuesEvent
|
2017-10-20 10:06:10
|
angular/angular-cli
|
https://api.github.com/repos/angular/angular-cli
|
closed
|
`ng test` doesn't watch when first compilation fails
|
command: test effort2: medium (days) freq3: high priority: 2 (required) severity3: broken workaround2: non-obvious
|
<!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [x] bug report -> please search issues before submitting
- [ ] feature request
```
### Versions.
<!--
Output from: `ng --version`.
If nothing, output from: `node --version` and `npm --version`.
Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?)
-->
@angular/cli: 1.4.5
node: 8.4.0
os: linux x64
@angular/animations: 4.4.4
@angular/common: 4.4.4
@angular/compiler: 4.4.4
@angular/core: 4.4.4
@angular/forms: 4.4.4
@angular/http: 4.4.4
@angular/platform-browser: 4.4.4
@angular/platform-browser-dynamic: 4.4.4
@angular/router: 4.4.4
@angular/cli: 1.4.5
@angular/compiler-cli: 4.4.4
@angular/language-service: 4.4.4
typescript: 2.3.4
### Repro steps.
<!--
Simple steps to reproduce this bug.
Please include: commands run, packages added, related code changes.
A link to a sample repo would help too.
-->
- `ng new testproject`
- introduce a syntax error to `./src/app/app.component.ts`
- run `ng test` and see the error message
- fix the syntax error and save
- doesn't recompile, tests are not rerun
### The log given by the failure.
<!-- Normally this include a stack trace and some more information. -->
### Desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
The `watch` should work independently of success/failure of the first compilation.
### Mention any other details that might be useful.
<!-- Please include a link to the repo if this is related to an OSS project. -->
- happened after update 1.3.2 > 1.4.5
- the issue happens only when the first compilation fails. i.e. when the first compilation passes and the next one fails, the further behavior is as expected (still watching and recompiling)
|
1.0
|
`ng test` doesn't watch when first compilation fails - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [x] bug report -> please search issues before submitting
- [ ] feature request
```
### Versions.
<!--
Output from: `ng --version`.
If nothing, output from: `node --version` and `npm --version`.
Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?)
-->
@angular/cli: 1.4.5
node: 8.4.0
os: linux x64
@angular/animations: 4.4.4
@angular/common: 4.4.4
@angular/compiler: 4.4.4
@angular/core: 4.4.4
@angular/forms: 4.4.4
@angular/http: 4.4.4
@angular/platform-browser: 4.4.4
@angular/platform-browser-dynamic: 4.4.4
@angular/router: 4.4.4
@angular/cli: 1.4.5
@angular/compiler-cli: 4.4.4
@angular/language-service: 4.4.4
typescript: 2.3.4
### Repro steps.
<!--
Simple steps to reproduce this bug.
Please include: commands run, packages added, related code changes.
A link to a sample repo would help too.
-->
- `ng new testproject`
- introduce a syntax error to `./src/app/app.component.ts`
- run `ng test` and see the error message
- fix the syntax error and save
- doesn't recompile, tests are not rerun
### The log given by the failure.
<!-- Normally this include a stack trace and some more information. -->
### Desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
The `watch` should work independently of success/failure of the first compilation.
### Mention any other details that might be useful.
<!-- Please include a link to the repo if this is related to an OSS project. -->
- happened after update 1.3.2 > 1.4.5
- the issue happens only when the first compilation fails. i.e. when the first compilation passes and the next one fails, the further behavior is as expected (still watching and recompiling)
|
test
|
ng test doesn t watch when first compilation fails if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request versions output from ng version if nothing output from node version and npm version windows linux incl distribution macos el capitan sierra angular cli node os linux angular animations angular common angular compiler angular core angular forms angular http angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli angular language service typescript repro steps simple steps to reproduce this bug please include commands run packages added related code changes a link to a sample repo would help too ng new testproject introduce a syntax error to src app app component ts run ng test and see the error message fix the syntax error and save doesn t recompile tests are not rerun the log given by the failure desired functionality what would like to see implemented what is the usecase the watch should work independently of success failure of the first compilation mention any other details that might be useful happened after update the issue happens only when the first compilation fails i e when the first compilation passes and the next one fails the further behavior is as expected still watching and recompiling
| 1
|
723,928
| 24,912,743,541
|
IssuesEvent
|
2022-10-30 02:56:51
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
closed
|
To enable async logging mechanism for a specific logger
|
Type/Improvement Priority/High commitment WUM Component/APIM
|
### Describe your problem(s)
As per the current implementation of the APIM 3.1.0 server, we can not enable the async logging mechanism for a specific logger. It throws the below error when enabling async logging via the log4j2.properties configuration.
```
org.ops4j.pax.logging.pax-logging-api [log4j2] ERROR : Log4J2 configuration problem: com.lmax.disruptor.dsl.Disruptor.<init>(Lcom/lmax/disruptor/EventFactory;ILjava/util/concurrent/ThreadFactory;Lcom/lmax/disruptor/dsl/ProducerType;Lcom/lmax/disruptor/WaitStrategy;)V Ignored FQCN: org.apache.logging.log4j.spi.AbstractLogger
java.lang.NoSuchMethodError: com.lmax.disruptor.dsl.Disruptor.<init>(Lcom/lmax/disruptor/EventFactory;ILjava/util/concurrent/ThreadFactory;Lcom/lmax/disruptor/dsl/ProducerType;Lcom/lmax/disruptor/WaitStrategy;)V
at org.apache.logging.log4j.core.async.AsyncLoggerConfigDisruptor.start(AsyncLoggerConfigDisruptor.java:231)
at org.apache.logging.log4j.core.config.AbstractConfiguration.start(AbstractConfiguration.java:296)
at org.apache.logging.log4j.core.LoggerContext.setConfiguration(LoggerContext.java:579)
at org.apache.logging.log4j.core.LoggerContext.start(LoggerContext.java:285)
at org.ops4j.pax.logging.log4j2.internal.PaxLoggingServiceImpl.configureLog4J2(PaxLoggingServiceImpl.java:439)
at org.ops4j.pax.logging.log4j2.internal.PaxLoggingServiceImpl.updated(PaxLoggingServiceImpl.java:272)
at org.ops4j.pax.logging.log4j2.internal.PaxLoggingServiceImpl$1ManagedPaxLoggingService.updated(PaxLoggingServiceImpl.java:581)
at org.eclipse.equinox.internal.cm.ManagedServiceTracker$1.run(ManagedServiceTracker.java:279)
at org.eclipse.equinox.internal.cm.SerializedTaskQueue$1.run(SerializedTaskQueue.java:40)
```
### Describe your solution
As per the official log4j2 document [1] , to enable async logging, disruptor-3.3.4.jar or higher is required. Currently, the available disruptor bundle version in the APIM 3.1.0 is 3.3.2.wso2v1. Therefore we need to upgrade the disruptor bundle version and then we can enable the async logging for a specific logger by defining the logger type as below in the log4j2.properties configuration.
Please refer to the below sample log4j2 configurations to enable the async logging for the `synapse-wire` logger
```
logger.synapse-wire.type = AsyncLogger
logger.synapse-wire.name = org.apache.synapse.transport.http.wire
logger.synapse-wire.level = DEBUG
```
[1]: https://logging.apache.org/log4j/2.x/manual/async.html
|
1.0
|
To enable async logging mechanism for a specific logger - ### Describe your problem(s)
As per the current implementation of the APIM 3.1.0 server, we can not enable the async logging mechanism for a specific logger. It throws the below error when enabling async logging via the log4j2.properties configuration.
```
org.ops4j.pax.logging.pax-logging-api [log4j2] ERROR : Log4J2 configuration problem: com.lmax.disruptor.dsl.Disruptor.<init>(Lcom/lmax/disruptor/EventFactory;ILjava/util/concurrent/ThreadFactory;Lcom/lmax/disruptor/dsl/ProducerType;Lcom/lmax/disruptor/WaitStrategy;)V Ignored FQCN: org.apache.logging.log4j.spi.AbstractLogger
java.lang.NoSuchMethodError: com.lmax.disruptor.dsl.Disruptor.<init>(Lcom/lmax/disruptor/EventFactory;ILjava/util/concurrent/ThreadFactory;Lcom/lmax/disruptor/dsl/ProducerType;Lcom/lmax/disruptor/WaitStrategy;)V
at org.apache.logging.log4j.core.async.AsyncLoggerConfigDisruptor.start(AsyncLoggerConfigDisruptor.java:231)
at org.apache.logging.log4j.core.config.AbstractConfiguration.start(AbstractConfiguration.java:296)
at org.apache.logging.log4j.core.LoggerContext.setConfiguration(LoggerContext.java:579)
at org.apache.logging.log4j.core.LoggerContext.start(LoggerContext.java:285)
at org.ops4j.pax.logging.log4j2.internal.PaxLoggingServiceImpl.configureLog4J2(PaxLoggingServiceImpl.java:439)
at org.ops4j.pax.logging.log4j2.internal.PaxLoggingServiceImpl.updated(PaxLoggingServiceImpl.java:272)
at org.ops4j.pax.logging.log4j2.internal.PaxLoggingServiceImpl$1ManagedPaxLoggingService.updated(PaxLoggingServiceImpl.java:581)
at org.eclipse.equinox.internal.cm.ManagedServiceTracker$1.run(ManagedServiceTracker.java:279)
at org.eclipse.equinox.internal.cm.SerializedTaskQueue$1.run(SerializedTaskQueue.java:40)
```
### Describe your solution
As per the official log4j2 document [1] , to enable async logging, disruptor-3.3.4.jar or higher is required. Currently, the available disruptor bundle version in the APIM 3.1.0 is 3.3.2.wso2v1. Therefore we need to upgrade the disruptor bundle version and then we can enable the async logging for a specific logger by defining the logger type as below in the log4j2.properties configuration.
Please refer to the below sample log4j2 configurations to enable the async logging for the `synapse-wire` logger
```
logger.synapse-wire.type = AsyncLogger
logger.synapse-wire.name = org.apache.synapse.transport.http.wire
logger.synapse-wire.level = DEBUG
```
[1]: https://logging.apache.org/log4j/2.x/manual/async.html
|
non_test
|
to enable async logging mechanism for a specific logger describe your problem s as per the current implementation of the apim server we can not enable the async logging mechanism for a specific logger it throws the below error when enabling async logging via the properties configuration org pax logging pax logging api error configuration problem com lmax disruptor dsl disruptor lcom lmax disruptor eventfactory iljava util concurrent threadfactory lcom lmax disruptor dsl producertype lcom lmax disruptor waitstrategy v ignored fqcn org apache logging spi abstractlogger java lang nosuchmethoderror com lmax disruptor dsl disruptor lcom lmax disruptor eventfactory iljava util concurrent threadfactory lcom lmax disruptor dsl producertype lcom lmax disruptor waitstrategy v at org apache logging core async asyncloggerconfigdisruptor start asyncloggerconfigdisruptor java at org apache logging core config abstractconfiguration start abstractconfiguration java at org apache logging core loggercontext setconfiguration loggercontext java at org apache logging core loggercontext start loggercontext java at org pax logging internal paxloggingserviceimpl paxloggingserviceimpl java at org pax logging internal paxloggingserviceimpl updated paxloggingserviceimpl java at org pax logging internal paxloggingserviceimpl updated paxloggingserviceimpl java at org eclipse equinox internal cm managedservicetracker run managedservicetracker java at org eclipse equinox internal cm serializedtaskqueue run serializedtaskqueue java describe your solution as per the official document to enable async logging disruptor jar or higher is required currently the available disruptor bundle version in the apim is therefore we need to upgrade the disruptor bundle version and then we can enable the async logging for a specific logger by defining the logger type as below in the properties configuration please refer to the below sample configurations to enable the async logging for the synapse wire logger logger synapse wire type asynclogger logger synapse wire name org apache synapse transport http wire logger synapse wire level debug
| 0
|
338,712
| 30,316,165,526
|
IssuesEvent
|
2023-07-10 15:42:42
|
newrelic/node-newrelic
|
https://api.github.com/repos/newrelic/node-newrelic
|
closed
|
Replace any t.equals assertions with t.equal
|
enhancement dev:tests
|
Some tests use t.equals assertions, which are being deprecated in favor of t.equal. We should update agent tests accordingly.
|
1.0
|
Replace any t.equals assertions with t.equal - Some tests use t.equals assertions, which are being deprecated in favor of t.equal. We should update agent tests accordingly.
|
test
|
replace any t equals assertions with t equal some tests use t equals assertions which are being deprecated in favor of t equal we should update agent tests accordingly
| 1
|
24,471
| 4,085,824,312
|
IssuesEvent
|
2016-06-01 00:55:19
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Socket_ConnectAsyncDnsEndPoint_HostNotFound test failed with TryAgain error
|
1 - Up Next System.Net test bug
|
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/windows_nt_release_prtest/2364/consoleFull
```
16:18:38 System.Net.Sockets.Tests.DnsEndPointTest.Socket_ConnectAsyncDnsEndPoint_HostNotFound [FAIL]
16:18:38 SocketError: TryAgain
16:18:38 Expected: True
16:18:38 Actual: False
16:18:38 Stack Trace:
16:18:38 d:\j\workspace\windows_nt_re---5067c899\src\System.Net.Sockets.Legacy\tests\FunctionalTests\DnsEndPointTest.cs(348,0): at System.Net.Sockets.Tests.DnsEndPointTest.AssertHostNotFoundOrNoData(SocketAsyncEventArgs args)
16:18:38 d:\j\workspace\windows_nt_re---5067c899\src\System.Net.Sockets.Legacy\tests\FunctionalTests\DnsEndPointTest.cs(196,0): at System.Net.Sockets.Tests.DnsEndPointTest.Socket_ConnectAsyncDnsEndPoint_HostNotFound()
```
|
1.0
|
Socket_ConnectAsyncDnsEndPoint_HostNotFound test failed with TryAgain error - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/windows_nt_release_prtest/2364/consoleFull
```
16:18:38 System.Net.Sockets.Tests.DnsEndPointTest.Socket_ConnectAsyncDnsEndPoint_HostNotFound [FAIL]
16:18:38 SocketError: TryAgain
16:18:38 Expected: True
16:18:38 Actual: False
16:18:38 Stack Trace:
16:18:38 d:\j\workspace\windows_nt_re---5067c899\src\System.Net.Sockets.Legacy\tests\FunctionalTests\DnsEndPointTest.cs(348,0): at System.Net.Sockets.Tests.DnsEndPointTest.AssertHostNotFoundOrNoData(SocketAsyncEventArgs args)
16:18:38 d:\j\workspace\windows_nt_re---5067c899\src\System.Net.Sockets.Legacy\tests\FunctionalTests\DnsEndPointTest.cs(196,0): at System.Net.Sockets.Tests.DnsEndPointTest.Socket_ConnectAsyncDnsEndPoint_HostNotFound()
```
|
test
|
socket connectasyncdnsendpoint hostnotfound test failed with tryagain error system net sockets tests dnsendpointtest socket connectasyncdnsendpoint hostnotfound socketerror tryagain expected true actual false stack trace d j workspace windows nt re src system net sockets legacy tests functionaltests dnsendpointtest cs at system net sockets tests dnsendpointtest asserthostnotfoundornodata socketasynceventargs args d j workspace windows nt re src system net sockets legacy tests functionaltests dnsendpointtest cs at system net sockets tests dnsendpointtest socket connectasyncdnsendpoint hostnotfound
| 1
|
350,787
| 31,932,318,084
|
IssuesEvent
|
2023-09-19 08:16:28
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix jax_lax_operators.test_jax_erf
|
JAX Frontend Sub Task Failing Test
|
| | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix jax_lax_operators.test_jax_erf - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6181361297/job/16779199230"><img src=https://img.shields.io/badge/-failure-red></a>
|
test
|
fix jax lax operators test jax erf numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
| 1
|
117,434
| 11,946,149,616
|
IssuesEvent
|
2020-04-03 07:33:44
|
hypergraph-xyz/desktop
|
https://api.github.com/repos/hypergraph-xyz/desktop
|
closed
|
Update Desktop readme with "what & why" + screenshot(s)
|
documentation in progress
|
Liven up the `README.md` a bit with:
- [x] Better explanation of what Hypergraph Desktop does
- [x] Better explanation of why Hypergraph Desktop exists
- [x] Screenshots of Hypergraph Desktop
|
1.0
|
Update Desktop readme with "what & why" + screenshot(s) - Liven up the `README.md` a bit with:
- [x] Better explanation of what Hypergraph Desktop does
- [x] Better explanation of why Hypergraph Desktop exists
- [x] Screenshots of Hypergraph Desktop
|
non_test
|
update desktop readme with what why screenshot s liven up the readme md a bit with better explanation of what hypergraph desktop does better explanation of why hypergraph desktop exists screenshots of hypergraph desktop
| 0
|
539,338
| 15,787,088,685
|
IssuesEvent
|
2021-04-01 18:41:58
|
code4romania/de-urgenta-android
|
https://api.github.com/repos/code4romania/de-urgenta-android
|
opened
|
Implement onboarding landing page
|
feature-onboarding high-priority :fire:
|
Implement the onboarding landing page as a fragment in the Onboarding Activity.
Please define related colors and text & button styles in the corresponding values files.
Some basic scaffolding was added, but it needs to be cleaned up.

[Link to design in Figma](https://www.figma.com/file/d6n3MPVRWbL5jP2NmvW96b/De_Urgenta?node-id=1466%3A530)
|
1.0
|
Implement onboarding landing page - Implement the onboarding landing page as a fragment in the Onboarding Activity.
Please define related colors and text & button styles in the corresponding values files.
Some basic scaffolding was added, but it needs to be cleaned up.

[Link to design in Figma](https://www.figma.com/file/d6n3MPVRWbL5jP2NmvW96b/De_Urgenta?node-id=1466%3A530)
|
non_test
|
implement onboarding landing page implement the onboarding landing page as a fragment in the onboarding activity please define related colors and text button styles in the corresponding values files some basic scaffolding was added but it needs to be cleaned up
| 0
|
120,259
| 10,111,670,853
|
IssuesEvent
|
2019-07-30 13:15:28
|
CodaProtocol/coda
|
https://api.github.com/repos/CodaProtocol/coda
|
closed
|
Time-as-a-functor throughout coda_main etc, s.t. we can run integration tests with fake time
|
planning-after-testnet
|
see `time-exposed2` branch for partial work there (has probably bit-rotted enough that it's worth starting over maybe?)
|
1.0
|
Time-as-a-functor throughout coda_main etc, s.t. we can run integration tests with fake time - see `time-exposed2` branch for partial work there (has probably bit-rotted enough that it's worth starting over maybe?)
|
test
|
time as a functor throughout coda main etc s t we can run integration tests with fake time see time branch for partial work there has probably bit rotted enough that it s worth starting over maybe
| 1
|
263,649
| 23,072,659,501
|
IssuesEvent
|
2022-07-25 19:41:10
|
kubernetes/test-infra
|
https://api.github.com/repos/kubernetes/test-infra
|
closed
|
TimeSeries "kubernetes.io/internal/addons/fluentbit/input_records_total" could not be written
|
kind/bug sig/testing lifecycle/rotten kind/oncall-hotlist needs-triage
|
At around 22:40 PST yesterday (Feb 24) there was a spike of these kinds of errors in k8s-prow:
```
{"caller":"exporterhelper/queued_retry.go:254", "error":"rpc error: code = InvalidArgument desc = One or more TimeSeries could not be written: Metric kind for metric kubernetes.io/internal/addons/fluentbit/input_records_total must be CUMULATIVE, but is GAUGE.: timeSeries[188-199]
error details: name = Unknown desc = total_point_count:200 success_point_count:188 errors:{status:{code:3} point_count:12}", "kind":"exporter", "level":"error", "msg":"Exporting failed. Try enabling retry_on_failure config option.", "name":"googlecloud", "stacktrace":"go.opentelemetry.io/collector/exporter/exporterhelper.(*retrySender).send
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/queued_retry.go:254
go.opentelemetry.io/collector/exporter/exporterhelper.(*metricsSenderWithObservability).send
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/metrics.go:114
go.opentelemetry.io/collector/exporter/exporterhelper.(*queuedRetrySender).send
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/queued_retry.go:174
go.opentelemetry.io/collector/exporter/exporterhelper.NewMetricsExporter.func2
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/metrics.go:94
go.opentelemetry.io/collector/consumer/consumerhelper.ConsumeMetricsFunc.ConsumeMetrics
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/consumer/consumerhelper/metrics.go:29
go.opentelemetry.io/collector/processor/batchprocessor.(*batchMetrics).export
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:288
go.opentelemetry.io/collector/processor/batchprocessor.(*batchProcessor).sendItems
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:184
go.opentelemetry.io/collector/processor/batchprocessor.(*batchProcessor).processItem
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:157
go.opentelemetry.io/collector/processor/batchprocessor.(*batchProcessor).startProcessingCycle
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:142", "ts":1.6457280215036135E9}
```
From this behavior I'm not sure if there's a misconfigured metric that we can fix ("One or more TimeSeries could not be written: Metric kind for metric kubernetes.io/internal/addons/fluentbit/input_records_total must be CUMULATIVE, but is GAUGE") or if, since this happened in a burst and hasn't reoccurred since, this is some temporary thing that requires no action.
|
1.0
|
TimeSeries "kubernetes.io/internal/addons/fluentbit/input_records_total" could not be written - At around 22:40 PST yesterday (Feb 24) there was a spike of these kinds of errors in k8s-prow:
```
{"caller":"exporterhelper/queued_retry.go:254", "error":"rpc error: code = InvalidArgument desc = One or more TimeSeries could not be written: Metric kind for metric kubernetes.io/internal/addons/fluentbit/input_records_total must be CUMULATIVE, but is GAUGE.: timeSeries[188-199]
error details: name = Unknown desc = total_point_count:200 success_point_count:188 errors:{status:{code:3} point_count:12}", "kind":"exporter", "level":"error", "msg":"Exporting failed. Try enabling retry_on_failure config option.", "name":"googlecloud", "stacktrace":"go.opentelemetry.io/collector/exporter/exporterhelper.(*retrySender).send
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/queued_retry.go:254
go.opentelemetry.io/collector/exporter/exporterhelper.(*metricsSenderWithObservability).send
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/metrics.go:114
go.opentelemetry.io/collector/exporter/exporterhelper.(*queuedRetrySender).send
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/queued_retry.go:174
go.opentelemetry.io/collector/exporter/exporterhelper.NewMetricsExporter.func2
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/exporter/exporterhelper/metrics.go:94
go.opentelemetry.io/collector/consumer/consumerhelper.ConsumeMetricsFunc.ConsumeMetrics
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/consumer/consumerhelper/metrics.go:29
go.opentelemetry.io/collector/processor/batchprocessor.(*batchMetrics).export
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:288
go.opentelemetry.io/collector/processor/batchprocessor.(*batchProcessor).sendItems
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:184
go.opentelemetry.io/collector/processor/batchprocessor.(*batchProcessor).processItem
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:157
go.opentelemetry.io/collector/processor/batchprocessor.(*batchProcessor).startProcessingCycle
/workspace/louhi_ws/gke-logmon/gke-metrics-agent/vendor/go.opentelemetry.io/collector/processor/batchprocessor/batch_processor.go:142", "ts":1.6457280215036135E9}
```
From this behavior I'm not sure if there's a misconfigured metric that we can fix ("One or more TimeSeries could not be written: Metric kind for metric kubernetes.io/internal/addons/fluentbit/input_records_total must be CUMULATIVE, but is GAUGE") or if, since this happened in a burst and hasn't reoccurred since, this is some temporary thing that requires no action.
|
test
|
timeseries kubernetes io internal addons fluentbit input records total could not be written at around pst yesterday feb there was a spike of these kinds of errors in prow caller exporterhelper queued retry go error rpc error code invalidargument desc one or more timeseries could not be written metric kind for metric kubernetes io internal addons fluentbit input records total must be cumulative but is gauge timeseries error details name unknown desc total point count success point count errors status code point count kind exporter level error msg exporting failed try enabling retry on failure config option name googlecloud stacktrace go opentelemetry io collector exporter exporterhelper retrysender send workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector exporter exporterhelper queued retry go go opentelemetry io collector exporter exporterhelper metricssenderwithobservability send workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector exporter exporterhelper metrics go go opentelemetry io collector exporter exporterhelper queuedretrysender send workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector exporter exporterhelper queued retry go go opentelemetry io collector exporter exporterhelper newmetricsexporter workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector exporter exporterhelper metrics go go opentelemetry io collector consumer consumerhelper consumemetricsfunc consumemetrics workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector consumer consumerhelper metrics go go opentelemetry io collector processor batchprocessor batchmetrics export workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector processor batchprocessor batch processor go go opentelemetry io collector processor batchprocessor batchprocessor senditems workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector processor batchprocessor batch processor go go opentelemetry io collector processor batchprocessor batchprocessor processitem workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector processor batchprocessor batch processor go go opentelemetry io collector processor batchprocessor batchprocessor startprocessingcycle workspace louhi ws gke logmon gke metrics agent vendor go opentelemetry io collector processor batchprocessor batch processor go ts from this behavior i m not sure if there s a misconfigured metric that we can fix one or more timeseries could not be written metric kind for metric kubernetes io internal addons fluentbit input records total must be cumulative but is gauge or if since this happened in a burst and hasn t reoccurred since this is some temporary thing that requires no action
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.