Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
31,452 | 2,732,910,563 | IssuesEvent | 2015-04-17 10:10:13 | tiku01/oryx-editor | https://api.github.com/repos/tiku01/oryx-editor | closed | Layout Management | auto-migrated Priority-Low Schedule-LongTerm Type-Enhancement | ```
The current layouting possibilities are very rudimentary. The goal is to
offer process model layout manager. Because the layouts of different
stencil sets differ, layouts must be stencil set specific.
```
Original issue reported on code.google.com by `NicoPete...@gmail.com` on 17 Apr 2008 at 1:44 | 1.0 | Layout Management - ```
The current layouting possibilities are very rudimentary. The goal is to
offer process model layout manager. Because the layouts of different
stencil sets differ, layouts must be stencil set specific.
```
Original issue reported on code.google.com by `NicoPete...@gmail.com` on 17 Apr 2008 at 1:44 | priority | layout management the current layouting possibilities are very rudimentary the goal is to offer process model layout manager because the layouts of different stencil sets differ layouts must be stencil set specific original issue reported on code google com by nicopete gmail com on apr at | 1 |
125,718 | 10,350,528,324 | IssuesEvent | 2019-09-05 03:10:50 | bskinn/stdio-mgr | https://api.github.com/repos/bskinn/stdio-mgr | reopened | Testing against console libraries | needs tests | colorama < 0.4 is incompatible, as its `StreamWrapper` only gained context manager features in 0.4 https://github.com/tartley/colorama/commit/2f4b564a6586709db2fed20e5d1f5856b1a482d2 and it pushes instances of its `StreamWrapper` into `sys.std*`
We could potentially work around this by detecting and patching its objects, but it probably isnt worth the effort. coala was forced to use < 0.4 because of `radon`, but they have bumped their min dependency, so we will too. Not many projects will be still wanting colorama < 0.4.
So this is probably best addressed with documentation of knownn incompatibilities.
However, it would be worth checking out other console libraries, and creating tests for each. | 1.0 | Testing against console libraries - colorama < 0.4 is incompatible, as its `StreamWrapper` only gained context manager features in 0.4 https://github.com/tartley/colorama/commit/2f4b564a6586709db2fed20e5d1f5856b1a482d2 and it pushes instances of its `StreamWrapper` into `sys.std*`
We could potentially work around this by detecting and patching its objects, but it probably isnt worth the effort. coala was forced to use < 0.4 because of `radon`, but they have bumped their min dependency, so we will too. Not many projects will be still wanting colorama < 0.4.
So this is probably best addressed with documentation of knownn incompatibilities.
However, it would be worth checking out other console libraries, and creating tests for each. | non_priority | testing against console libraries colorama is incompatible as its streamwrapper only gained context manager features in and it pushes instances of its streamwrapper into sys std we could potentially work around this by detecting and patching its objects but it probably isnt worth the effort coala was forced to use because of radon but they have bumped their min dependency so we will too not many projects will be still wanting colorama so this is probably best addressed with documentation of knownn incompatibilities however it would be worth checking out other console libraries and creating tests for each | 0 |
480,272 | 13,838,963,448 | IssuesEvent | 2020-10-14 07:11:48 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.wenxuecity.com - site is not usable | browser-firefox engine-gecko priority-normal | <!-- @browser: Firefox Nightly 83 for Android -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4260.0 Safari/537.36 Edg/87.0.637.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/59824 -->
**URL**: https://www.wenxuecity.com
**Browser / Version**: Firefox Nightly 83 for Android
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
I get a "Secure Connection Failed" message since Firefox does not trust the website certificate for www.wenxuecity.com. However, Chrome for Android (as well as desktop Chrome) loads www.wenxuecity.com with no problems/issues. This is only a problem with Firefox.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/e0f82e6c-e547-4528-a0b1-afcfaf70154a.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | 1.0 | www.wenxuecity.com - site is not usable - <!-- @browser: Firefox Nightly 83 for Android -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4260.0 Safari/537.36 Edg/87.0.637.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/59824 -->
**URL**: https://www.wenxuecity.com
**Browser / Version**: Firefox Nightly 83 for Android
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
I get a "Secure Connection Failed" message since Firefox does not trust the website certificate for www.wenxuecity.com. However, Chrome for Android (as well as desktop Chrome) loads www.wenxuecity.com with no problems/issues. This is only a problem with Firefox.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/e0f82e6c-e547-4528-a0b1-afcfaf70154a.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | priority | site is not usable url browser version firefox nightly for android operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce i get a secure connection failed message since firefox does not trust the website certificate for however chrome for android as well as desktop chrome loads with no problems issues this is only a problem with firefox view the screenshot img alt screenshot src browser configuration none from with โค๏ธ | 1 |
139,159 | 11,253,478,467 | IssuesEvent | 2020-01-11 16:33:38 | kiwicom/schemathesis | https://api.github.com/repos/kiwicom/schemathesis | closed | Fix flaky tests | Priority: Low Type: Testing | There are tests in the generation package that sometimes fail - we need to disable some hypothesis healthchecks | 1.0 | Fix flaky tests - There are tests in the generation package that sometimes fail - we need to disable some hypothesis healthchecks | non_priority | fix flaky tests there are tests in the generation package that sometimes fail we need to disable some hypothesis healthchecks | 0 |
139,248 | 18,849,675,349 | IssuesEvent | 2021-11-11 19:06:05 | RG4421/multi-juicer | https://api.github.com/repos/RG4421/multi-juicer | opened | CVE-2021-3757 (High) detected in immer-8.0.1.tgz | security vulnerability | ## CVE-2021-3757 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-8.0.1.tgz</b></p></summary>
<p>Create your next immutable state by mutating the current one</p>
<p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-8.0.1.tgz">https://registry.npmjs.org/immer/-/immer-8.0.1.tgz</a></p>
<p>Path to dependency file: multi-juicer/juice-balancer/ui/package.json</p>
<p>Path to vulnerable library: multi-juicer/juice-balancer/ui/node_modules/immer/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- :x: **immer-8.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/multi-juicer/commit/8576a853ef65531c776b5ea4aac618580e1b0354">8576a853ef65531c776b5ea4aac618580e1b0354</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
immer is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-09-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3757>CVE-2021-3757</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/23d38099-71cd-42ed-a77a-71e68094adfa/">https://huntr.dev/bounties/23d38099-71cd-42ed-a77a-71e68094adfa/</a></p>
<p>Release Date: 2021-09-02</p>
<p>Fix Resolution: immer - 9.0.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"immer","packageVersion":"8.0.1","packageFilePaths":["/juice-balancer/ui/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:4.0.3;react-dev-utils:11.0.4;immer:8.0.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"immer - 9.0.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3757","vulnerabilityDetails":"immer is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3757","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-3757 (High) detected in immer-8.0.1.tgz - ## CVE-2021-3757 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-8.0.1.tgz</b></p></summary>
<p>Create your next immutable state by mutating the current one</p>
<p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-8.0.1.tgz">https://registry.npmjs.org/immer/-/immer-8.0.1.tgz</a></p>
<p>Path to dependency file: multi-juicer/juice-balancer/ui/package.json</p>
<p>Path to vulnerable library: multi-juicer/juice-balancer/ui/node_modules/immer/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- :x: **immer-8.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/multi-juicer/commit/8576a853ef65531c776b5ea4aac618580e1b0354">8576a853ef65531c776b5ea4aac618580e1b0354</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
immer is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-09-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3757>CVE-2021-3757</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/23d38099-71cd-42ed-a77a-71e68094adfa/">https://huntr.dev/bounties/23d38099-71cd-42ed-a77a-71e68094adfa/</a></p>
<p>Release Date: 2021-09-02</p>
<p>Fix Resolution: immer - 9.0.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"immer","packageVersion":"8.0.1","packageFilePaths":["/juice-balancer/ui/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:4.0.3;react-dev-utils:11.0.4;immer:8.0.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"immer - 9.0.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3757","vulnerabilityDetails":"immer is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3757","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in immer tgz cve high severity vulnerability vulnerable library immer tgz create your next immutable state by mutating the current one library home page a href path to dependency file multi juicer juice balancer ui package json path to vulnerable library multi juicer juice balancer ui node modules immer package json dependency hierarchy react scripts tgz root library react dev utils tgz x immer tgz vulnerable library found in head commit a href found in base branch master vulnerability details immer is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution immer isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree react scripts react dev utils immer isminimumfixversionavailable true minimumfixversion immer basebranches vulnerabilityidentifier cve vulnerabilitydetails immer is vulnerable to improperly controlled modification of object prototype attributes pollution vulnerabilityurl | 0 |
239,205 | 7,787,298,738 | IssuesEvent | 2018-06-06 21:54:27 | coq/coq | https://api.github.com/repos/coq/coq | opened | Interrupting printing completely borks coqtop | kind: anomaly part: STM priority: high | #### Version
8.8.0
#### Operating system
Linux
#### Description of the problem
Consider the code
```coq
Require Import Coq.ZArith.ZArith.
Definition foo := Eval vm_compute in List.repeat Z.div_eucl 20.
Set Printing Depth 1000000.
Set Printing All.
Print foo.
```
In ProofGeneral: Execute up to the `Set Printing All.` Then execute the `Print foo.`, wait a second or two or three (until the `*coq*` buffer starts showing output). Then spam `C-c` to send `SIGINT` to Coq. Then execute any command, such as `Check 1.` or `Reset Initial.` or even just something silly like `R.`. I get
```coq
Error:
Currently, the parsing api only supports parsing at the tip of the document.
You wanted to parse at: 1 but the current tip is: 63
```
cc @ejgallego @gares | 1.0 | Interrupting printing completely borks coqtop - #### Version
8.8.0
#### Operating system
Linux
#### Description of the problem
Consider the code
```coq
Require Import Coq.ZArith.ZArith.
Definition foo := Eval vm_compute in List.repeat Z.div_eucl 20.
Set Printing Depth 1000000.
Set Printing All.
Print foo.
```
In ProofGeneral: Execute up to the `Set Printing All.` Then execute the `Print foo.`, wait a second or two or three (until the `*coq*` buffer starts showing output). Then spam `C-c` to send `SIGINT` to Coq. Then execute any command, such as `Check 1.` or `Reset Initial.` or even just something silly like `R.`. I get
```coq
Error:
Currently, the parsing api only supports parsing at the tip of the document.
You wanted to parse at: 1 but the current tip is: 63
```
cc @ejgallego @gares | priority | interrupting printing completely borks coqtop version operating system linux description of the problem consider the code coq require import coq zarith zarith definition foo eval vm compute in list repeat z div eucl set printing depth set printing all print foo in proofgeneral execute up to the set printing all then execute the print foo wait a second or two or three until the coq buffer starts showing output then spam c c to send sigint to coq then execute any command such as check or reset initial or even just something silly like r i get coq error currently the parsing api only supports parsing at the tip of the document you wanted to parse at but the current tip is cc ejgallego gares | 1 |
776,154 | 27,248,878,140 | IssuesEvent | 2023-02-22 05:58:16 | italanta/elewa | https://api.github.com/repos/italanta/elewa | closed | ๐งฑ ๐จ Keyword Jump Block | enhancement good first issue priority | ## User stories
As a user I want a block that saves keywords from the users response.
---
Quick links: [Summary](#summary) | [Description](#description) | [Requirements](#template)
---
### Summary
1. As a user I want to be able to save keywords from the user using a keyword jump block to determine what my response will be.
### Description
The Keyword jump block is used to receive the a keyword from a response and then determine which block or connection will be made next.
It has the following features:
```markdown
1. A header (Inherited from the block component)
2. An input that contains the default route
```
### Requirements:
```markdown
1. All the block functionality still works i.e copy, delete, add option and rendering of the block.
2. The component is pixel perfect compared to the design flow of the bot.
```
| 1.0 | ๐งฑ ๐จ Keyword Jump Block - ## User stories
As a user I want a block that saves keywords from the users response.
---
Quick links: [Summary](#summary) | [Description](#description) | [Requirements](#template)
---
### Summary
1. As a user I want to be able to save keywords from the user using a keyword jump block to determine what my response will be.
### Description
The Keyword jump block is used to receive the a keyword from a response and then determine which block or connection will be made next.
It has the following features:
```markdown
1. A header (Inherited from the block component)
2. An input that contains the default route
```
### Requirements:
```markdown
1. All the block functionality still works i.e copy, delete, add option and rendering of the block.
2. The component is pixel perfect compared to the design flow of the bot.
```
| priority | ๐งฑ ๐จ keyword jump block user stories as a user i want a block that saves keywords from the users response quick links summary description template summary as a user i want to be able to save keywords from the user using a keyword jump block to determine what my response will be description the keyword jump block is used to receive the a keyword from a response and then determine which block or connection will be made next it has the following features markdown a header inherited from the block component an input that contains the default route requirements markdown all the block functionality still works i e copy delete add option and rendering of the block the component is pixel perfect compared to the design flow of the bot | 1 |
479,597 | 13,803,672,502 | IssuesEvent | 2020-10-11 04:34:14 | The-Mu-Foundation/Mutorials | https://api.github.com/repos/The-Mu-Foundation/Mutorials | reopened | Dynamic problems | feature_request high_priority | This should randomise the numbers in each problem, and calculate the answer choices again, while respecting significant figures.
| 1.0 | Dynamic problems - This should randomise the numbers in each problem, and calculate the answer choices again, while respecting significant figures.
| priority | dynamic problems this should randomise the numbers in each problem and calculate the answer choices again while respecting significant figures | 1 |
783,877 | 27,549,888,865 | IssuesEvent | 2023-03-07 14:18:48 | unfoldingWord/gateway-edit | https://api.github.com/repos/unfoldingWord/gateway-edit | closed | Make the feedback form a draggable modal so the user can view their work environment when reporting feedback | Priority/Medium enhancement QA/Passed | Issue:
---
Currently, the feedback form is a fixed page. It should be a draggable feedback form.
DoD:
---
The feedback form is moveable.
Details:
---
v0.9.0 build 87451c8
The the user's project cannot be seen when submitting feedback. In many cases the feedback will be related to what is in their workspace so it would be easier for them to provide accurate feedback if they could see it when filling out the form. | 1.0 | Make the feedback form a draggable modal so the user can view their work environment when reporting feedback - Issue:
---
Currently, the feedback form is a fixed page. It should be a draggable feedback form.
DoD:
---
The feedback form is moveable.
Details:
---
v0.9.0 build 87451c8
The the user's project cannot be seen when submitting feedback. In many cases the feedback will be related to what is in their workspace so it would be easier for them to provide accurate feedback if they could see it when filling out the form. | priority | make the feedback form a draggable modal so the user can view their work environment when reporting feedback issue currently the feedback form is a fixed page it should be a draggable feedback form dod the feedback form is moveable details build the the user s project cannot be seen when submitting feedback in many cases the feedback will be related to what is in their workspace so it would be easier for them to provide accurate feedback if they could see it when filling out the form | 1 |
183,755 | 31,761,102,980 | IssuesEvent | 2023-09-12 05:16:16 | prgrms-fe-devcourse/FEDC4_TMI_HOMERS_OFF | https://api.github.com/repos/prgrms-fe-devcourse/FEDC4_TMI_HOMERS_OFF | closed | design: Subscribe, Tab, TabItem ๊ณตํต ์ปดํฌ๋ํธ | ๐จ Design | ## ์ถ๊ฐํ๋ ค๋ ๊ธฐ๋ฅ์ด ์ด๋ ํ ๋ฌธ์ ํน์ ๊ธฐ๋ฅ๊ณผ ์ฐ๊ด๋์ด ์๋์?
ํ๋ก์ ํธ ๊ณตํต ์ปดํฌ๋ํธ์ธ ์๋ ์ปดํฌ๋ํธ๋ค์ ์ถ๊ฐํ๋ ค๊ณ ํฉ๋๋ค.
- Subscribe
- Tab
## ์ํ๋ ๊ธฐ๋ฅ ์ถ๊ฐ
์ถ๊ฐํ๋ ค๋ ๊ธฐ๋ฅ์ ๋ช
ํํ๊ณ ๊ฐ๊ฒฐํ๊ฒ ์ค๋ช
ํด์ฃผ์ธ์
- [ ] Subscribe ์ปดํฌ๋ํธ
- [ ] Tab ์ปดํฌ๋ํธ
## ์๊ฐ ์ค์ธ ๊ธฐ๋ฅ ์ถ๊ฐ ๋ฐฉ์
### Subscribe ์ปดํฌ๋ํธ
- ์์ props
- Subscribers
- Subscribings
### Tab ์ปดํฌ๋ํธ
- ์์ props
- isActive
- children (TabItem)
### TabItem ์ปดํฌ๋ํธ
- ์์ props
- title
- isActive
ํด๊ฒฐ์ฑ
์ผ๋ก ๊ฐ๋จํ๊ฒ ์๊ฐํ ๊ธฐ๋ฅ์ ๋ฐฉํฅ ํน์ ์ปดํฌ๋ํธ๋ฅผ ์ค๋ช
ํด์ฃผ์ธ์
## ETC
์คํฌ๋ฆฐ์ท์ด๋ ๊ธฐ๋ฅ ๋ฑ ์ถ๊ฐ ์๋ฃ๋ฅผ ๊ธฐ์ ํด ์ฃผ์ธ์
### Tab ์ปดํฌ๋ํธ

### Subscribe ์ปดํฌ๋ํธ

| 1.0 | design: Subscribe, Tab, TabItem ๊ณตํต ์ปดํฌ๋ํธ - ## ์ถ๊ฐํ๋ ค๋ ๊ธฐ๋ฅ์ด ์ด๋ ํ ๋ฌธ์ ํน์ ๊ธฐ๋ฅ๊ณผ ์ฐ๊ด๋์ด ์๋์?
ํ๋ก์ ํธ ๊ณตํต ์ปดํฌ๋ํธ์ธ ์๋ ์ปดํฌ๋ํธ๋ค์ ์ถ๊ฐํ๋ ค๊ณ ํฉ๋๋ค.
- Subscribe
- Tab
## ์ํ๋ ๊ธฐ๋ฅ ์ถ๊ฐ
์ถ๊ฐํ๋ ค๋ ๊ธฐ๋ฅ์ ๋ช
ํํ๊ณ ๊ฐ๊ฒฐํ๊ฒ ์ค๋ช
ํด์ฃผ์ธ์
- [ ] Subscribe ์ปดํฌ๋ํธ
- [ ] Tab ์ปดํฌ๋ํธ
## ์๊ฐ ์ค์ธ ๊ธฐ๋ฅ ์ถ๊ฐ ๋ฐฉ์
### Subscribe ์ปดํฌ๋ํธ
- ์์ props
- Subscribers
- Subscribings
### Tab ์ปดํฌ๋ํธ
- ์์ props
- isActive
- children (TabItem)
### TabItem ์ปดํฌ๋ํธ
- ์์ props
- title
- isActive
ํด๊ฒฐ์ฑ
์ผ๋ก ๊ฐ๋จํ๊ฒ ์๊ฐํ ๊ธฐ๋ฅ์ ๋ฐฉํฅ ํน์ ์ปดํฌ๋ํธ๋ฅผ ์ค๋ช
ํด์ฃผ์ธ์
## ETC
์คํฌ๋ฆฐ์ท์ด๋ ๊ธฐ๋ฅ ๋ฑ ์ถ๊ฐ ์๋ฃ๋ฅผ ๊ธฐ์ ํด ์ฃผ์ธ์
### Tab ์ปดํฌ๋ํธ

### Subscribe ์ปดํฌ๋ํธ

| non_priority | design subscribe tab tabitem ๊ณตํต ์ปดํฌ๋ํธ ์ถ๊ฐํ๋ ค๋ ๊ธฐ๋ฅ์ด ์ด๋ ํ ๋ฌธ์ ํน์ ๊ธฐ๋ฅ๊ณผ ์ฐ๊ด๋์ด ์๋์ ํ๋ก์ ํธ ๊ณตํต ์ปดํฌ๋ํธ์ธ ์๋ ์ปดํฌ๋ํธ๋ค์ ์ถ๊ฐํ๋ ค๊ณ ํฉ๋๋ค subscribe tab ์ํ๋ ๊ธฐ๋ฅ ์ถ๊ฐ ์ถ๊ฐํ๋ ค๋ ๊ธฐ๋ฅ์ ๋ช
ํํ๊ณ ๊ฐ๊ฒฐํ๊ฒ ์ค๋ช
ํด์ฃผ์ธ์ subscribe ์ปดํฌ๋ํธ tab ์ปดํฌ๋ํธ ์๊ฐ ์ค์ธ ๊ธฐ๋ฅ ์ถ๊ฐ ๋ฐฉ์ subscribe ์ปดํฌ๋ํธ ์์ props subscribers subscribings tab ์ปดํฌ๋ํธ ์์ props isactive children tabitem tabitem ์ปดํฌ๋ํธ ์์ props title isactive ํด๊ฒฐ์ฑ
์ผ๋ก ๊ฐ๋จํ๊ฒ ์๊ฐํ ๊ธฐ๋ฅ์ ๋ฐฉํฅ ํน์ ์ปดํฌ๋ํธ๋ฅผ ์ค๋ช
ํด์ฃผ์ธ์ etc ์คํฌ๋ฆฐ์ท์ด๋ ๊ธฐ๋ฅ ๋ฑ ์ถ๊ฐ ์๋ฃ๋ฅผ ๊ธฐ์ ํด ์ฃผ์ธ์ tab ์ปดํฌ๋ํธ subscribe ์ปดํฌ๋ํธ | 0 |
25,273 | 24,933,660,422 | IssuesEvent | 2022-10-31 13:38:55 | informalsystems/apalache | https://api.github.com/repos/informalsystems/apalache | closed | Add types to output JSON traces/counterexamples | usability feature | Currently, `violation.json` or `example.json` files produced by Apalache lack any type information. Consequently, these traces cannot be imported via `JsonToTlaViaBuilder`. We should construct these counterexamples in Apalache using the typed builder in model recovery, so they are automatically typed. | True | Add types to output JSON traces/counterexamples - Currently, `violation.json` or `example.json` files produced by Apalache lack any type information. Consequently, these traces cannot be imported via `JsonToTlaViaBuilder`. We should construct these counterexamples in Apalache using the typed builder in model recovery, so they are automatically typed. | non_priority | add types to output json traces counterexamples currently violation json or example json files produced by apalache lack any type information consequently these traces cannot be imported via jsontotlaviabuilder we should construct these counterexamples in apalache using the typed builder in model recovery so they are automatically typed | 0 |
6,051 | 5,259,281,696 | IssuesEvent | 2017-02-03 03:01:24 | sshnet/SSH.NET | https://api.github.com/repos/sshnet/SSH.NET | closed | SharpSSH run faster than SSH.NET | performance | Hi, This is my test results, SharpSSH run each command takes about 20 milliseconds, SSH.NET run each command takes about 100 milliseconds.
Is it possible to improve performance?Thanks!
| True | SharpSSH run faster than SSH.NET - Hi, This is my test results, SharpSSH run each command takes about 20 milliseconds, SSH.NET run each command takes about 100 milliseconds.
Is it possible to improve performance?Thanks!
| non_priority | sharpssh run faster than ssh net hi this is my test results sharpssh run each command takes about milliseconds ssh net run each command takes about milliseconds is it possible to improve performance thanks | 0 |
186,157 | 14,394,659,172 | IssuesEvent | 2020-12-03 01:49:17 | github-vet/rangeclosure-findings | https://api.github.com/repos/github-vet/rangeclosure-findings | closed | axetang/AxeCoding: Golang/gopkg/_gopkgsrc/net/interface_test.go; 24 LoC | fresh small test |
Found a possible issue in [axetang/AxeCoding](https://www.github.com/axetang/AxeCoding) at [Golang/gopkg/_gopkgsrc/net/interface_test.go](https://github.com/axetang/AxeCoding/blob/0acec6a9480856a2438844261cccaf80513ca7a1/Golang/gopkg/_gopkgsrc/net/interface_test.go#L58-L81)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to ifi at line 69 may start a goroutine
[Click here to see the code in its original context.](https://github.com/axetang/AxeCoding/blob/0acec6a9480856a2438844261cccaf80513ca7a1/Golang/gopkg/_gopkgsrc/net/interface_test.go#L58-L81)
<details>
<summary>Click here to show the 24 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ifi := range ift {
ifxi, err := InterfaceByIndex(ifi.Index)
if err != nil {
t.Fatal(err)
}
switch runtime.GOOS {
case "solaris":
if ifxi.Index != ifi.Index {
t.Errorf("got %v; want %v", ifxi, ifi)
}
default:
if !reflect.DeepEqual(ifxi, &ifi) {
t.Errorf("got %v; want %v", ifxi, ifi)
}
}
ifxn, err := InterfaceByName(ifi.Name)
if err != nil {
t.Fatal(err)
}
if !reflect.DeepEqual(ifxn, &ifi) {
t.Errorf("got %v; want %v", ifxn, ifi)
}
t.Logf("%s: flags=%v index=%d mtu=%d hwaddr=%v", ifi.Name, ifi.Flags, ifi.Index, ifi.MTU, ifi.HardwareAddr)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 0acec6a9480856a2438844261cccaf80513ca7a1
| 1.0 | axetang/AxeCoding: Golang/gopkg/_gopkgsrc/net/interface_test.go; 24 LoC -
Found a possible issue in [axetang/AxeCoding](https://www.github.com/axetang/AxeCoding) at [Golang/gopkg/_gopkgsrc/net/interface_test.go](https://github.com/axetang/AxeCoding/blob/0acec6a9480856a2438844261cccaf80513ca7a1/Golang/gopkg/_gopkgsrc/net/interface_test.go#L58-L81)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to ifi at line 69 may start a goroutine
[Click here to see the code in its original context.](https://github.com/axetang/AxeCoding/blob/0acec6a9480856a2438844261cccaf80513ca7a1/Golang/gopkg/_gopkgsrc/net/interface_test.go#L58-L81)
<details>
<summary>Click here to show the 24 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ifi := range ift {
ifxi, err := InterfaceByIndex(ifi.Index)
if err != nil {
t.Fatal(err)
}
switch runtime.GOOS {
case "solaris":
if ifxi.Index != ifi.Index {
t.Errorf("got %v; want %v", ifxi, ifi)
}
default:
if !reflect.DeepEqual(ifxi, &ifi) {
t.Errorf("got %v; want %v", ifxi, ifi)
}
}
ifxn, err := InterfaceByName(ifi.Name)
if err != nil {
t.Fatal(err)
}
if !reflect.DeepEqual(ifxn, &ifi) {
t.Errorf("got %v; want %v", ifxn, ifi)
}
t.Logf("%s: flags=%v index=%d mtu=%d hwaddr=%v", ifi.Name, ifi.Flags, ifi.Index, ifi.MTU, ifi.HardwareAddr)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 0acec6a9480856a2438844261cccaf80513ca7a1
| non_priority | axetang axecoding golang gopkg gopkgsrc net interface test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to ifi at line may start a goroutine click here to show the line s of go which triggered the analyzer go for ifi range ift ifxi err interfacebyindex ifi index if err nil t fatal err switch runtime goos case solaris if ifxi index ifi index t errorf got v want v ifxi ifi default if reflect deepequal ifxi ifi t errorf got v want v ifxi ifi ifxn err interfacebyname ifi name if err nil t fatal err if reflect deepequal ifxn ifi t errorf got v want v ifxn ifi t logf s flags v index d mtu d hwaddr v ifi name ifi flags ifi index ifi mtu ifi hardwareaddr leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
682,103 | 23,332,794,981 | IssuesEvent | 2022-08-09 07:19:03 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | twitter.com - design is broken | browser-firefox priority-critical engine-gecko | <!-- @browser: Firefox 105.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:105.0) Gecko/20100101 Firefox/105.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/108604 -->
**URL**: https://twitter.com/
**Browser / Version**: Firefox 105.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Items not fully visible
**Steps to Reproduce**:
in chrome ok;
in firefox not open page correct...
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/3ef0a0de-6f0a-4eb8-9d3c-dd60bf9302a1.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220806094424</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/8/f3142578-ba32-43e4-b537-678212cd5cbb)
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | 1.0 | twitter.com - design is broken - <!-- @browser: Firefox 105.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:105.0) Gecko/20100101 Firefox/105.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/108604 -->
**URL**: https://twitter.com/
**Browser / Version**: Firefox 105.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Items not fully visible
**Steps to Reproduce**:
in chrome ok;
in firefox not open page correct...
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/3ef0a0de-6f0a-4eb8-9d3c-dd60bf9302a1.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220806094424</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/8/f3142578-ba32-43e4-b537-678212cd5cbb)
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | priority | twitter com design is broken url browser version firefox operating system windows tested another browser yes chrome problem type design is broken description items not fully visible steps to reproduce in chrome ok in firefox not open page correct view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with โค๏ธ | 1 |
313,605 | 23,484,318,066 | IssuesEvent | 2022-08-17 13:18:17 | valentingavran/anglify | https://api.github.com/repos/valentingavran/anglify | closed | docs-bug: Component APIs | documentation | ### Documentation Feedback
Developers who want to use Anglify have to find the available input properties and slots from the samples or analyze the code.
Instead, we should list all input properties and slots of the respective components directly on the web page in a table.
### Affected documentation page
all component pages | 1.0 | docs-bug: Component APIs - ### Documentation Feedback
Developers who want to use Anglify have to find the available input properties and slots from the samples or analyze the code.
Instead, we should list all input properties and slots of the respective components directly on the web page in a table.
### Affected documentation page
all component pages | non_priority | docs bug component apis documentation feedback developers who want to use anglify have to find the available input properties and slots from the samples or analyze the code instead we should list all input properties and slots of the respective components directly on the web page in a table affected documentation page all component pages | 0 |
819,801 | 30,751,538,286 | IssuesEvent | 2023-07-28 19:46:19 | AdguardTeam/CoreLibs | https://api.github.com/repos/AdguardTeam/CoreLibs | closed | Why Has The development of the macOS App Stopped? | Priority: P4 | Why has the development of the macOS Adguard system app being ignored?
There have been NO substantial improvements for quite a long time.
The development of Adguard macOS app has seemingly stagnated.
The development of the macOS app NOT been updated to provide features which are equal to or better than the Extension.
I fully acknowledge the macOS system app is far more technically capable than any Extension.
So why is the macOS app user interface is lacking features that the Extension contains and why is the macOS app is so very badly designed and that it is need of a major overhaul.
The Technical Issue of the Filtering Logs NOT recording Cosmetic/Custom User Rules has been known by Adguard for at least 5 years and they done absolutely NOTHING to fix this.
The folks at Adguard seem to have NO INTEREST in fixing this.
The fact is that a Filtering Log which is incompetent and lacks features creates issues when items on a webpage are blocked and you have NO RECORDS TO SHOW WHAT COSMETIC ITEMS ARE BLOCKED, This makes it a very difficult and time consuming task to find out what is blocking parts of a webpage.
People whom are paying for a License to use a paid product deserve much better.
It is absolutely disgusting and abhorrent that the staff at Adguard have been so very, very, lazy that they can not be bothered improving their paid for Adguard and introducing what it sorely lacks and bringing this product up to its full potential.
Adguard is not improving, development has seeming ground to a halt and the problem with the Filtering Log has been present for 5 very long years now.
The customers of Adguard are paying hard earned cash for this product and in return we get a very badly designed product.
Paying customers deserve much better treatment, this is simply not good enough | 1.0 | Why Has The development of the macOS App Stopped? - Why has the development of the macOS Adguard system app being ignored?
There have been NO substantial improvements for quite a long time.
The development of Adguard macOS app has seemingly stagnated.
The development of the macOS app NOT been updated to provide features which are equal to or better than the Extension.
I fully acknowledge the macOS system app is far more technically capable than any Extension.
So why is the macOS app user interface is lacking features that the Extension contains and why is the macOS app is so very badly designed and that it is need of a major overhaul.
The Technical Issue of the Filtering Logs NOT recording Cosmetic/Custom User Rules has been known by Adguard for at least 5 years and they done absolutely NOTHING to fix this.
The folks at Adguard seem to have NO INTEREST in fixing this.
The fact is that a Filtering Log which is incompetent and lacks features creates issues when items on a webpage are blocked and you have NO RECORDS TO SHOW WHAT COSMETIC ITEMS ARE BLOCKED, This makes it a very difficult and time consuming task to find out what is blocking parts of a webpage.
People whom are paying for a License to use a paid product deserve much better.
It is absolutely disgusting and abhorrent that the staff at Adguard have been so very, very, lazy that they can not be bothered improving their paid for Adguard and introducing what it sorely lacks and bringing this product up to its full potential.
Adguard is not improving, development has seeming ground to a halt and the problem with the Filtering Log has been present for 5 very long years now.
The customers of Adguard are paying hard earned cash for this product and in return we get a very badly designed product.
Paying customers deserve much better treatment, this is simply not good enough | priority | why has the development of the macos app stopped why has the development of the macos adguard system app being ignored there have been no substantial improvements for quite a long time the development of adguard macos app has seemingly stagnated the development of the macos app not been updated to provide features which are equal to or better than the extension i fully acknowledge the macos system app is far more technically capable than any extension so why is the macos app user interface is lacking features that the extension contains and why is the macos app is so very badly designed and that it is need of a major overhaul the technical issue of the filtering logs not recording cosmetic custom user rules has been known by adguard for at least years and they done absolutely nothing to fix this the folks at adguard seem to have no interest in fixing this the fact is that a filtering log which is incompetent and lacks features creates issues when items on a webpage are blocked and you have no records to show what cosmetic items are blocked this makes it a very difficult and time consuming task to find out what is blocking parts of a webpage people whom are paying for a license to use a paid product deserve much better it is absolutely disgusting and abhorrent that the staff at adguard have been so very very lazy that they can not be bothered improving their paid for adguard and introducing what it sorely lacks and bringing this product up to its full potential adguard is not improving development has seeming ground to a halt and the problem with the filtering log has been present for very long years now the customers of adguard are paying hard earned cash for this product and in return we get a very badly designed product paying customers deserve much better treatment this is simply not good enough | 1 |
155,500 | 24,472,052,793 | IssuesEvent | 2022-10-07 21:10:02 | OpenLiberty/liberty-tools-vscode | https://api.github.com/repos/OpenLiberty/liberty-tools-vscode | closed | Allow for projects to be manually added/removed from the Liberty dashboard | enhancement 5 high priority design GUI SVT-req | Eg, right-click -> Add to Liberty dashboard
Only projects manually added or removed to the dashboard should persist between VS Code sessions | 1.0 | Allow for projects to be manually added/removed from the Liberty dashboard - Eg, right-click -> Add to Liberty dashboard
Only projects manually added or removed to the dashboard should persist between VS Code sessions | non_priority | allow for projects to be manually added removed from the liberty dashboard eg right click add to liberty dashboard only projects manually added or removed to the dashboard should persist between vs code sessions | 0 |
500,522 | 14,501,331,017 | IssuesEvent | 2020-12-11 19:21:11 | GoogleCloudPlatform/cloud-code-vscode | https://api.github.com/repos/GoogleCloudPlatform/cloud-code-vscode | closed | Could not login to Google Cloud SDK in Google Cloud Shell | kind/bug priority/p1 | **Environment**: Extension version: 1.7.0 VSCode version: 1.51.1 OS: linux **Description**: **Repro step**:
- Follow instruction at:
https://cloud.google.com/code/docs/vscode/quickstart-remote-dev#node.js
- Stuck at: Click here to login to Google
- It's seem like the extension could not understand gcloud config in remote development Google Cloud Shell | 1.0 | Could not login to Google Cloud SDK in Google Cloud Shell - **Environment**: Extension version: 1.7.0 VSCode version: 1.51.1 OS: linux **Description**: **Repro step**:
- Follow instruction at:
https://cloud.google.com/code/docs/vscode/quickstart-remote-dev#node.js
- Stuck at: Click here to login to Google
- It's seem like the extension could not understand gcloud config in remote development Google Cloud Shell | priority | could not login to google cloud sdk in google cloud shell environment extension version vscode version os linux description repro step follow instruction at stuck at click here to login to google it s seem like the extension could not understand gcloud config in remote development google cloud shell | 1 |
610,212 | 18,900,908,222 | IssuesEvent | 2021-11-16 00:43:06 | internetarchive/openlibrary | https://api.github.com/repos/internetarchive/openlibrary | opened | Automate steps to move solr from solrbuilder server to production server | Affects: Admin/Maintenance Type: Feature Request Priority: 2 Lead: @cdrini | This process is annoying and finnicky; getting automated would save a good chunk of time and reduce risk.
- [x] Run steps manually (done in #5493 )
- [x] Write out as bash, and run each step-by-step (done in #5502)
- [ ] Open draft PR
- [ ] Try running the entire script! Do as part of #5827 | 1.0 | Automate steps to move solr from solrbuilder server to production server - This process is annoying and finnicky; getting automated would save a good chunk of time and reduce risk.
- [x] Run steps manually (done in #5493 )
- [x] Write out as bash, and run each step-by-step (done in #5502)
- [ ] Open draft PR
- [ ] Try running the entire script! Do as part of #5827 | priority | automate steps to move solr from solrbuilder server to production server this process is annoying and finnicky getting automated would save a good chunk of time and reduce risk run steps manually done in write out as bash and run each step by step done in open draft pr try running the entire script do as part of | 1 |
226,716 | 24,996,511,976 | IssuesEvent | 2022-11-03 01:10:33 | hzhaoCS/juice-shop | https://api.github.com/repos/hzhaoCS/juice-shop | opened | socket.io-2.1.13.tgz: 1 vulnerabilities (highest severity is: 9.8) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-2.1.13.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/socket.io-parser/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/hzhaoCS/juice-shop/commit/8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe">8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (socket.io version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-2421](https://www.mend.io/vulnerability-database/CVE-2022-2421) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | socket.io-parser-4.1.2.tgz | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2421</summary>
### Vulnerable Library - <b>socket.io-parser-4.1.2.tgz</b></p>
<p>socket.io protocol parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-4.1.2.tgz">https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-4.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/socket.io-parser/package.json</p>
<p>
Dependency Hierarchy:
- socket.io-2.1.13.tgz (Root Library)
- socket.io-parser-3.0.0.tgz
- :x: **socket.io-parser-4.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hzhaoCS/juice-shop/commit/8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe">8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Due to improper type validation in attachment parsing the Socket.io js library, it is possible to overwrite the _placeholder object which allows an attacker to place references to functions at arbitrary places in the resulting query object.
<p>Publish Date: 2022-10-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-2421>CVE-2022-2421</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://csirt.divd.nl/cases/DIVD-2022-00045/">https://csirt.divd.nl/cases/DIVD-2022-00045/</a></p>
<p>Release Date: 2022-10-26</p>
<p>Fix Resolution: socket.io-parser - 4.0.5,4.2.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | socket.io-2.1.13.tgz: 1 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-2.1.13.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/socket.io-parser/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/hzhaoCS/juice-shop/commit/8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe">8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (socket.io version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-2421](https://www.mend.io/vulnerability-database/CVE-2022-2421) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | socket.io-parser-4.1.2.tgz | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2421</summary>
### Vulnerable Library - <b>socket.io-parser-4.1.2.tgz</b></p>
<p>socket.io protocol parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-4.1.2.tgz">https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-4.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/socket.io-parser/package.json</p>
<p>
Dependency Hierarchy:
- socket.io-2.1.13.tgz (Root Library)
- socket.io-parser-3.0.0.tgz
- :x: **socket.io-parser-4.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hzhaoCS/juice-shop/commit/8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe">8c9b0228ffbc4bbd54967e5ee1dcb66ebde454fe</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Due to improper type validation in attachment parsing the Socket.io js library, it is possible to overwrite the _placeholder object which allows an attacker to place references to functions at arbitrary places in the resulting query object.
<p>Publish Date: 2022-10-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-2421>CVE-2022-2421</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://csirt.divd.nl/cases/DIVD-2022-00045/">https://csirt.divd.nl/cases/DIVD-2022-00045/</a></p>
<p>Release Date: 2022-10-26</p>
<p>Fix Resolution: socket.io-parser - 4.0.5,4.2.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_priority | socket io tgz vulnerabilities highest severity is vulnerable library socket io tgz path to dependency file package json path to vulnerable library node modules socket io parser package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in socket io version remediation available high socket io parser tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library socket io parser tgz socket io protocol parser library home page a href path to dependency file package json path to vulnerable library node modules socket io parser package json dependency hierarchy socket io tgz root library socket io parser tgz x socket io parser tgz vulnerable library found in head commit a href found in base branch master vulnerability details due to improper type validation in attachment parsing the socket io js library it is possible to overwrite the placeholder object which allows an attacker to place references to functions at arbitrary places in the resulting query object publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution socket io parser step up your open source security game with mend | 0 |
133,352 | 29,048,271,915 | IssuesEvent | 2023-05-13 21:24:51 | OpenModelica/OpenModelica | https://api.github.com/repos/OpenModelica/OpenModelica | opened | Codegen fails to handle record assignments in when equations | COMP/OMC/Codegen | Please run the following MWE:
```modelica
model TestRecordAssignment
Complex v;
discrete Complex v0;
equation
v = Complex(sin(time), cos(time));
when time > 0.5 then
v0 = pre(v);
end when;
end TestRecordAssignment;
```
The Codegen phase fails with
```
[5] 23:21:40 Translation Error
[CodegenC.tpl: 6665:11-6665:11]: Template error: No runtime support for this record assignment: v0 = pre(v).
```
This feature is needed to fix a problem in the PowerGrids library.
@mahge can you please give it a look? | 1.0 | Codegen fails to handle record assignments in when equations - Please run the following MWE:
```modelica
model TestRecordAssignment
Complex v;
discrete Complex v0;
equation
v = Complex(sin(time), cos(time));
when time > 0.5 then
v0 = pre(v);
end when;
end TestRecordAssignment;
```
The Codegen phase fails with
```
[5] 23:21:40 Translation Error
[CodegenC.tpl: 6665:11-6665:11]: Template error: No runtime support for this record assignment: v0 = pre(v).
```
This feature is needed to fix a problem in the PowerGrids library.
@mahge can you please give it a look? | non_priority | codegen fails to handle record assignments in when equations please run the following mwe modelica model testrecordassignment complex v discrete complex equation v complex sin time cos time when time then pre v end when end testrecordassignment the codegen phase fails with translation error template error no runtime support for this record assignment pre v this feature is needed to fix a problem in the powergrids library mahge can you please give it a look | 0 |
324,465 | 24,000,369,836 | IssuesEvent | 2022-09-14 10:56:12 | coq/platform | https://api.github.com/repos/coq/platform | closed | Pick readme files: include OCaml version in header | part: documentation kind: enhancement | As discussed in Zulip, the auto generated pick readme files should contain the OCaml version / flags in the header. | 1.0 | Pick readme files: include OCaml version in header - As discussed in Zulip, the auto generated pick readme files should contain the OCaml version / flags in the header. | non_priority | pick readme files include ocaml version in header as discussed in zulip the auto generated pick readme files should contain the ocaml version flags in the header | 0 |
37,185 | 8,228,225,421 | IssuesEvent | 2018-09-07 03:50:56 | OmniSharp/omnisharp-vscode | https://api.github.com/repos/OmniSharp/omnisharp-vscode | reopened | Asking Omnisharp to add a namespace results in the change being made in a new editor pane | Editor-Code Actions Needs More Info | ## Environment data
`dotnet --info` output:
```
.NET Core SDK (reflecting any global.json):
Version: 2.1.302
Commit: 9048955601
Runtime Environment:
OS Name: Windows
OS Version: 10.0.17134
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\2.1.302\
Host (useful for support):
Version: 2.1.2
Commit: 811c3ce6c0
.NET Core SDKs installed:
2.1.302 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
VS Code version: 1.26.1
C# Extension version: 1.15.2
## Steps to reproduce
1. Refer to a type in a namespace not yet referenced in a `using` statement
2. Open the "lightbulb" and select "using The.Namespace"
## Expected behavior
Change is made to the file in the current editor pane
## Actual behavior
New editor opens for same file, where the using statement has been added | 1.0 | Asking Omnisharp to add a namespace results in the change being made in a new editor pane - ## Environment data
`dotnet --info` output:
```
.NET Core SDK (reflecting any global.json):
Version: 2.1.302
Commit: 9048955601
Runtime Environment:
OS Name: Windows
OS Version: 10.0.17134
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\2.1.302\
Host (useful for support):
Version: 2.1.2
Commit: 811c3ce6c0
.NET Core SDKs installed:
2.1.302 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
VS Code version: 1.26.1
C# Extension version: 1.15.2
## Steps to reproduce
1. Refer to a type in a namespace not yet referenced in a `using` statement
2. Open the "lightbulb" and select "using The.Namespace"
## Expected behavior
Change is made to the file in the current editor pane
## Actual behavior
New editor opens for same file, where the using statement has been added | non_priority | asking omnisharp to add a namespace results in the change being made in a new editor pane environment data dotnet info output net core sdk reflecting any global json version commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk host useful for support version commit net core sdks installed net core runtimes installed microsoft aspnetcore all microsoft aspnetcore app microsoft netcore app to install additional net core runtimes or sdks vs code version c extension version steps to reproduce refer to a type in a namespace not yet referenced in a using statement open the lightbulb and select using the namespace expected behavior change is made to the file in the current editor pane actual behavior new editor opens for same file where the using statement has been added | 0 |
491,192 | 14,146,644,219 | IssuesEvent | 2020-11-10 19:32:17 | powercomm/PCM-Dashboard | https://api.github.com/repos/powercomm/PCM-Dashboard | opened | When a new PCM is added, history chart is blank, user experience needs improvement | Deficiency Priority: Low | Type 30 Edit form needs to fill history chart when no history points are returned from PCM_AC, a straight line will make it easier to see the threshold alignments. | 1.0 | When a new PCM is added, history chart is blank, user experience needs improvement - Type 30 Edit form needs to fill history chart when no history points are returned from PCM_AC, a straight line will make it easier to see the threshold alignments. | priority | when a new pcm is added history chart is blank user experience needs improvement type edit form needs to fill history chart when no history points are returned from pcm ac a straight line will make it easier to see the threshold alignments | 1 |
792 | 2,515,131,161 | IssuesEvent | 2015-01-15 16:37:50 | gabrielfedel/mapaculturacampinas | https://api.github.com/repos/gabrielfedel/mapaculturacampinas | closed | Resultado da busca | design mediaup | 1. Estรก desconfigurado, precisamos que siga o padrรฃo:
* รcone, tรญtulo, descriรงรฃo

<br><br>
2. Necessรกrio customizar e embelezar o texto que aparece acima:
Resultado da pesquisa (XX-XX total), _padronizar o tamanho do texto_
<br><br>
3. Texto abaixo, alinhar layout, remover bullets

| 1.0 | Resultado da busca - 1. Estรก desconfigurado, precisamos que siga o padrรฃo:
* รcone, tรญtulo, descriรงรฃo

<br><br>
2. Necessรกrio customizar e embelezar o texto que aparece acima:
Resultado da pesquisa (XX-XX total), _padronizar o tamanho do texto_
<br><br>
3. Texto abaixo, alinhar layout, remover bullets

| non_priority | resultado da busca estรก desconfigurado precisamos que siga o padrรฃo รญcone tรญtulo descriรงรฃo necessรกrio customizar e embelezar o texto que aparece acima resultado da pesquisa xx xx total padronizar o tamanho do texto texto abaixo alinhar layout remover bullets | 0 |
493,989 | 14,243,036,089 | IssuesEvent | 2020-11-19 03:20:16 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | Update images that have never been launched as a container | Priority: Low Status: Available Status: Stale Type: Enhancement | I have a use case where I want images updated on a machine but I haven't necessarily ran the image yet as a container. In this case, `--include-stopped` takes no effect (a container never ran), and there are no options to look at images only.
Expanded problem:
I have a CI system. The worker instances, when created, pull images that we've approved to run. When a new worker starts up it uses the local version of the image (never reaches out to a repository for images). Our system will periodically recycle the worker instances and may load up older versions of images on them.
We'd like something in place that looks at the images, determines they are older, then updates them. This way when the first job comes into a machine it'll have the latest version of that image instead of the one we had specified when the image was built. Only certain images should be able to pull the "latest" while others are ignored and pinned to the version we specified.
**Describe the solution you'd like**
I would like a flag that allows watchtower to look at images, not containers, on a machine and determine if a new version needs to be pulled.
**Describe alternatives you've considered**
Given our scenario we could run all the images on startup then shut them down. The `--include-stopped` flag would likely work for us then.
**Additional context**
I'm willing to do the work. I was already looking at where the `--include-stopped` flag is implemented and am proficient with go. The biggest hurdle is that this relies on images and not containers. I want to know if this is something the community would want before I go putting in the effort to open a PR with this functionality. If it's not valuable, or is too specific to my use case, I will consider other workarounds or roll my own image update process. I'd like to build something more useful though :).
| 1.0 | Update images that have never been launched as a container - I have a use case where I want images updated on a machine but I haven't necessarily ran the image yet as a container. In this case, `--include-stopped` takes no effect (a container never ran), and there are no options to look at images only.
Expanded problem:
I have a CI system. The worker instances, when created, pull images that we've approved to run. When a new worker starts up it uses the local version of the image (never reaches out to a repository for images). Our system will periodically recycle the worker instances and may load up older versions of images on them.
We'd like something in place that looks at the images, determines they are older, then updates them. This way when the first job comes into a machine it'll have the latest version of that image instead of the one we had specified when the image was built. Only certain images should be able to pull the "latest" while others are ignored and pinned to the version we specified.
**Describe the solution you'd like**
I would like a flag that allows watchtower to look at images, not containers, on a machine and determine if a new version needs to be pulled.
**Describe alternatives you've considered**
Given our scenario we could run all the images on startup then shut them down. The `--include-stopped` flag would likely work for us then.
**Additional context**
I'm willing to do the work. I was already looking at where the `--include-stopped` flag is implemented and am proficient with go. The biggest hurdle is that this relies on images and not containers. I want to know if this is something the community would want before I go putting in the effort to open a PR with this functionality. If it's not valuable, or is too specific to my use case, I will consider other workarounds or roll my own image update process. I'd like to build something more useful though :).
| priority | update images that have never been launched as a container i have a use case where i want images updated on a machine but i haven t necessarily ran the image yet as a container in this case include stopped takes no effect a container never ran and there are no options to look at images only expanded problem i have a ci system the worker instances when created pull images that we ve approved to run when a new worker starts up it uses the local version of the image never reaches out to a repository for images our system will periodically recycle the worker instances and may load up older versions of images on them we d like something in place that looks at the images determines they are older then updates them this way when the first job comes into a machine it ll have the latest version of that image instead of the one we had specified when the image was built only certain images should be able to pull the latest while others are ignored and pinned to the version we specified describe the solution you d like i would like a flag that allows watchtower to look at images not containers on a machine and determine if a new version needs to be pulled describe alternatives you ve considered given our scenario we could run all the images on startup then shut them down the include stopped flag would likely work for us then additional context i m willing to do the work i was already looking at where the include stopped flag is implemented and am proficient with go the biggest hurdle is that this relies on images and not containers i want to know if this is something the community would want before i go putting in the effort to open a pr with this functionality if it s not valuable or is too specific to my use case i will consider other workarounds or roll my own image update process i d like to build something more useful though | 1 |
195,752 | 15,554,735,173 | IssuesEvent | 2021-03-16 04:34:57 | CheoR/pack-it-up | https://api.github.com/repos/CheoR/pack-it-up | opened | QR code has to be made after all other information is filled | DB QR bug documentation enhancement | Title
QR code has to be made after all other information is filled
Narrative
As a: n app, QRCode should only be created when there's an associated box and item information.
I want: QRcode generated using box/item information as seed
so that: every QR code i unique
Acceptance criteria
Given: user creates and satisfies box/item information
When: when user saves the inofrmation
Then: app creates a unique QR code based on data payload provided by box/items tables. | 1.0 | QR code has to be made after all other information is filled - Title
QR code has to be made after all other information is filled
Narrative
As a: n app, QRCode should only be created when there's an associated box and item information.
I want: QRcode generated using box/item information as seed
so that: every QR code i unique
Acceptance criteria
Given: user creates and satisfies box/item information
When: when user saves the inofrmation
Then: app creates a unique QR code based on data payload provided by box/items tables. | non_priority | qr code has to be made after all other information is filled title qr code has to be made after all other information is filled narrative as a n app qrcode should only be created when there s an associated box and item information i want qrcode generated using box item information as seed so that every qr code i unique acceptance criteria given user creates and satisfies box item information when when user saves the inofrmation then app creates a unique qr code based on data payload provided by box items tables | 0 |
60,762 | 17,023,514,674 | IssuesEvent | 2021-07-03 02:25:11 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | shift-click to join ways doesn't work anymore | Component: potlatch (flash editor) Priority: critical Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 12.49am, Monday, 23rd November 2009]**
I've tried this on Mac/Safari, Mac/FireFox, PC/Firefox and PC/IE.
When I select a way and then shift-left-click on a connected way to join then, Potlatch simply selects the second way without joining them.
It can be repro'd by simply creating two connected ways.
Marking critical, since this make Potlatch effectively useless for serious editing.
| 1.0 | shift-click to join ways doesn't work anymore - **[Submitted to the original trac issue database at 12.49am, Monday, 23rd November 2009]**
I've tried this on Mac/Safari, Mac/FireFox, PC/Firefox and PC/IE.
When I select a way and then shift-left-click on a connected way to join then, Potlatch simply selects the second way without joining them.
It can be repro'd by simply creating two connected ways.
Marking critical, since this make Potlatch effectively useless for serious editing.
| non_priority | shift click to join ways doesn t work anymore i ve tried this on mac safari mac firefox pc firefox and pc ie when i select a way and then shift left click on a connected way to join then potlatch simply selects the second way without joining them it can be repro d by simply creating two connected ways marking critical since this make potlatch effectively useless for serious editing | 0 |
669,105 | 22,612,303,659 | IssuesEvent | 2022-06-29 18:18:13 | vmware-tanzu/tanzu-framework | https://api.github.com/repos/vmware-tanzu/tanzu-framework | opened | Shallow copy of non-map fields within clusterbootstrapclone.go#addMissingFields() | priority/important-longterm area/addons kind/feature needs-triage | (This is used to request new product features)
**Describe the feature request**
With current implementation under `addons/pkg/util/clusterbootstrapclone/clusterbootstrapclone.go#addMissingFields`, it does a shallow copy of the non-map fields (`destination[keyInFrom] = valueInFrom`) . This works fine when handling non-pointer values, but we might want to consider primitive pointer cases. (e.g., *int)
**Describe alternatives you've considered**
NA
**Affected product area (please put an X in all that apply)**
- [ ] APIs
- [x] Addons
- [ ] CLI
- [ ] Docs
- [ ] IAM
- [ ] Installation
- [ ] Plugin
- [ ] Security
- [ ] Test and Release
- [ ] User Experience
**Additional context**
| 1.0 | Shallow copy of non-map fields within clusterbootstrapclone.go#addMissingFields() - (This is used to request new product features)
**Describe the feature request**
With current implementation under `addons/pkg/util/clusterbootstrapclone/clusterbootstrapclone.go#addMissingFields`, it does a shallow copy of the non-map fields (`destination[keyInFrom] = valueInFrom`) . This works fine when handling non-pointer values, but we might want to consider primitive pointer cases. (e.g., *int)
**Describe alternatives you've considered**
NA
**Affected product area (please put an X in all that apply)**
- [ ] APIs
- [x] Addons
- [ ] CLI
- [ ] Docs
- [ ] IAM
- [ ] Installation
- [ ] Plugin
- [ ] Security
- [ ] Test and Release
- [ ] User Experience
**Additional context**
| priority | shallow copy of non map fields within clusterbootstrapclone go addmissingfields this is used to request new product features describe the feature request with current implementation under addons pkg util clusterbootstrapclone clusterbootstrapclone go addmissingfields it does a shallow copy of the non map fields destination valueinfrom this works fine when handling non pointer values but we might want to consider primitive pointer cases e g int describe alternatives you ve considered na affected product area please put an x in all that apply apis addons cli docs iam installation plugin security test and release user experience additional context | 1 |
3,389 | 5,826,587,517 | IssuesEvent | 2017-05-08 05:43:09 | avant-js/hapi-api-example | https://api.github.com/repos/avant-js/hapi-api-example | closed | Create a simple Hapi server | requirement | Create a server with Hapi and an exemple route and a static files server | 1.0 | Create a simple Hapi server - Create a server with Hapi and an exemple route and a static files server | non_priority | create a simple hapi server create a server with hapi and an exemple route and a static files server | 0 |
109,555 | 4,389,451,239 | IssuesEvent | 2016-08-08 22:10:24 | gophish/gophish | https://api.github.com/repos/gophish/gophish | closed | Email templates and issues with attached files | bug med-priority | After I create an e-mail template with an attached phishing file of my own, I get two sort of issues when the campaign is launched:
1. With text only content and my file attached, it sends blank emails: no text at all, only the enclosed file is there.
2. With HTML, if I check the "Add tracking image", it does not send any e-mail.
In fact, only the case of HTML + attached file without tracking image works as expected.
I don't get any error message of any kind. | 1.0 | Email templates and issues with attached files - After I create an e-mail template with an attached phishing file of my own, I get two sort of issues when the campaign is launched:
1. With text only content and my file attached, it sends blank emails: no text at all, only the enclosed file is there.
2. With HTML, if I check the "Add tracking image", it does not send any e-mail.
In fact, only the case of HTML + attached file without tracking image works as expected.
I don't get any error message of any kind. | priority | email templates and issues with attached files after i create an e mail template with an attached phishing file of my own i get two sort of issues when the campaign is launched with text only content and my file attached it sends blank emails no text at all only the enclosed file is there with html if i check the add tracking image it does not send any e mail in fact only the case of html attached file without tracking image works as expected i don t get any error message of any kind | 1 |
573,191 | 17,023,606,572 | IssuesEvent | 2021-07-03 02:53:28 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | aeroway=rc_runway | Component: mapnik Priority: minor Resolution: wontfix Type: enhancement | **[Submitted to the original trac issue database at 5.12am, Friday, 18th June 2010]**
Runways for radio-controlled aircraft should be rendered similar width and zoom levels as residential streets. | 1.0 | aeroway=rc_runway - **[Submitted to the original trac issue database at 5.12am, Friday, 18th June 2010]**
Runways for radio-controlled aircraft should be rendered similar width and zoom levels as residential streets. | priority | aeroway rc runway runways for radio controlled aircraft should be rendered similar width and zoom levels as residential streets | 1 |
318,081 | 27,283,987,880 | IssuesEvent | 2023-02-23 12:10:51 | helmholtz-analytics/heat | https://api.github.com/repos/helmholtz-analytics/heat | closed | Set up GitHub Actions | testing organizational | **Feature functionality**
GitHub introduced the GitHub Actions.
This is a CI/CD pipeline similar to the one GitLab already provides.
The benefit is that each build can be split into smaller more readable separated workflow units.
Added to that, it's also possible to use self-hosted runners. This can be used to run our tests on a private machine with GPU support.
For more information check out the feature description [here](https://github.com/features/actions).
Tasks:
- [ ] Move current build process to GitHub actions
- [ ] Split build process into smaller units that can run in parallel
- [ ] Create deploy workflow for publishing the library on PyPi
- [x] Set up a self-hosted runner with GPU (HDF machine in Jรผlich)
- [x] Create test stage that also tests on the GPU | 1.0 | Set up GitHub Actions - **Feature functionality**
GitHub introduced the GitHub Actions.
This is a CI/CD pipeline similar to the one GitLab already provides.
The benefit is that each build can be split into smaller more readable separated workflow units.
Added to that, it's also possible to use self-hosted runners. This can be used to run our tests on a private machine with GPU support.
For more information check out the feature description [here](https://github.com/features/actions).
Tasks:
- [ ] Move current build process to GitHub actions
- [ ] Split build process into smaller units that can run in parallel
- [ ] Create deploy workflow for publishing the library on PyPi
- [x] Set up a self-hosted runner with GPU (HDF machine in Jรผlich)
- [x] Create test stage that also tests on the GPU | non_priority | set up github actions feature functionality github introduced the github actions this is a ci cd pipeline similar to the one gitlab already provides the benefit is that each build can be split into smaller more readable separated workflow units added to that it s also possible to use self hosted runners this can be used to run our tests on a private machine with gpu support for more information check out the feature description tasks move current build process to github actions split build process into smaller units that can run in parallel create deploy workflow for publishing the library on pypi set up a self hosted runner with gpu hdf machine in jรผlich create test stage that also tests on the gpu | 0 |
287,671 | 31,856,176,923 | IssuesEvent | 2023-09-15 07:37:19 | Trinadh465/linux-4.1.15_CVE-2023-26607 | https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-26607 | opened | CVE-2023-0458 (Medium) detected in linuxlinux-4.6 | Mend: dependency security vulnerability | ## CVE-2023-0458 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A speculative pointer dereference problem exists in the Linux Kernel on the do_prlimit() function. The resource argument value is controlled and is used in pointer arithmetic for the 'rlim' variable and can be used to leak the contents. We recommend upgrading past version 6.1.8 or commitย 739790605705ddcf18f21782b9c99ad7d53a8c11
<p>Publish Date: 2023-04-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0458>CVE-2023-0458</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0458">https://www.linuxkernelcves.com/cves/CVE-2023-0458</a></p>
<p>Release Date: 2023-04-26</p>
<p>Fix Resolution: v4.14.304,v4.19.271,v5.4.230,v5.10.165,v5.15.90,v6.1.8,v6.2-rc5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-0458 (Medium) detected in linuxlinux-4.6 - ## CVE-2023-0458 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A speculative pointer dereference problem exists in the Linux Kernel on the do_prlimit() function. The resource argument value is controlled and is used in pointer arithmetic for the 'rlim' variable and can be used to leak the contents. We recommend upgrading past version 6.1.8 or commitย 739790605705ddcf18f21782b9c99ad7d53a8c11
<p>Publish Date: 2023-04-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0458>CVE-2023-0458</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0458">https://www.linuxkernelcves.com/cves/CVE-2023-0458</a></p>
<p>Release Date: 2023-04-26</p>
<p>Fix Resolution: v4.14.304,v4.19.271,v5.4.230,v5.10.165,v5.15.90,v6.1.8,v6.2-rc5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a speculative pointer dereference problem exists in the linux kernel on the do prlimit function the resource argument value is controlled and is used in pointer arithmetic for the rlim variable and can be used to leak the contents we recommend upgrading past version or commitย publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
13,733 | 3,770,957,061 | IssuesEvent | 2016-03-16 16:07:25 | pvlib/pvlib-python | https://api.github.com/repos/pvlib/pvlib-python | closed | add zenodo doi to readme, add preferred citation(s) to readme and docs | documentation Release | @Calama-Consulting and I discussed creating DOIs for pvlib releases. I just hooked up our repo to [Zenodo](https://zenodo.org) so that it will automatically create DOIs for each release. I'll add the badge to the Readme after we release 0.2.
To test the system, I rereleased the [PVSC 2015 repo](https://github.com/pvlib/pvsc2015) to trigger the DOI creation and then I added the new badge. It takes a little bit of time for DOIs to propagate, but hopefully the link on that readme will work soon.
We should also discuss how we would want people to cite pvlib in academic papers. See e.g.
http://ipython.org/citing.html
http://matplotlib.org/citing.html
http://deeplearning.net/software/theano/citation.html#citation
My only recommendation for now is to just use the papers in the PVSC 2014 and 2015 proceedings. I don't know how to specifically cite github contributors. | 1.0 | add zenodo doi to readme, add preferred citation(s) to readme and docs - @Calama-Consulting and I discussed creating DOIs for pvlib releases. I just hooked up our repo to [Zenodo](https://zenodo.org) so that it will automatically create DOIs for each release. I'll add the badge to the Readme after we release 0.2.
To test the system, I rereleased the [PVSC 2015 repo](https://github.com/pvlib/pvsc2015) to trigger the DOI creation and then I added the new badge. It takes a little bit of time for DOIs to propagate, but hopefully the link on that readme will work soon.
We should also discuss how we would want people to cite pvlib in academic papers. See e.g.
http://ipython.org/citing.html
http://matplotlib.org/citing.html
http://deeplearning.net/software/theano/citation.html#citation
My only recommendation for now is to just use the papers in the PVSC 2014 and 2015 proceedings. I don't know how to specifically cite github contributors. | non_priority | add zenodo doi to readme add preferred citation s to readme and docs calama consulting and i discussed creating dois for pvlib releases i just hooked up our repo to so that it will automatically create dois for each release i ll add the badge to the readme after we release to test the system i rereleased the to trigger the doi creation and then i added the new badge it takes a little bit of time for dois to propagate but hopefully the link on that readme will work soon we should also discuss how we would want people to cite pvlib in academic papers see e g my only recommendation for now is to just use the papers in the pvsc and proceedings i don t know how to specifically cite github contributors | 0 |
191,760 | 15,303,246,662 | IssuesEvent | 2021-02-24 15:35:48 | AllianceBioversityCIAT/onecgiar-submission-tool | https://api.github.com/repos/AllianceBioversityCIAT/onecgiar-submission-tool | closed | [FE - SG] (QA platform 2021) Users manual and guidance: Assessors | documentation enhancement | An user manual for assessor must be provided. This manual should explain the expected behavior and assessment of the assessors in the QA platform.
It also should describe new features added in 2021.
- [ ] 2020 assessment features explanation
- [ ] 2020 assessment features and updates explanation
**Deliverable:** Assessor user manual. | 1.0 | [FE - SG] (QA platform 2021) Users manual and guidance: Assessors - An user manual for assessor must be provided. This manual should explain the expected behavior and assessment of the assessors in the QA platform.
It also should describe new features added in 2021.
- [ ] 2020 assessment features explanation
- [ ] 2020 assessment features and updates explanation
**Deliverable:** Assessor user manual. | non_priority | qa platform users manual and guidance assessors an user manual for assessor must be provided this manual should explain the expected behavior and assessment of the assessors in the qa platform it also should describe new features added in assessment features explanation assessment features and updates explanation deliverable assessor user manual | 0 |
466,505 | 13,425,584,728 | IssuesEvent | 2020-09-06 10:51:32 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 develop-15] Civics: capitol's overlapping cases | Category: UI Priority: Medium Status: Reopen | Need some adjustments here and there
1) Article

2) Removed Article
This one looks ugly, actually.

| 1.0 | [0.9.0 develop-15] Civics: capitol's overlapping cases - Need some adjustments here and there
1) Article

2) Removed Article
This one looks ugly, actually.

| priority | civics capitol s overlapping cases need some adjustments here and there article removed article this one looks ugly actually | 1 |
294,015 | 9,012,132,993 | IssuesEvent | 2019-02-05 16:12:33 | Microsoft/Recommenders | https://api.github.com/repos/Microsoft/Recommenders | closed | [BUG] generate conda file is returning an error | bug high priority | ### *What* is affected by this bug?
generate conda file is returning an error
### In *which* platform does it happen?
* *Azure Data Science Virtual Machine.*
### *How* do we replicate the issue?
```
sh scripts/generate_conda_file.sh
scripts/generate_conda_file.sh: 64: scripts/generate_conda_file.sh: Syntax error: "(" unexpected (expecting "then")
```
| 1.0 | [BUG] generate conda file is returning an error - ### *What* is affected by this bug?
generate conda file is returning an error
### In *which* platform does it happen?
* *Azure Data Science Virtual Machine.*
### *How* do we replicate the issue?
```
sh scripts/generate_conda_file.sh
scripts/generate_conda_file.sh: 64: scripts/generate_conda_file.sh: Syntax error: "(" unexpected (expecting "then")
```
| priority | generate conda file is returning an error what is affected by this bug generate conda file is returning an error in which platform does it happen azure data science virtual machine how do we replicate the issue sh scripts generate conda file sh scripts generate conda file sh scripts generate conda file sh syntax error unexpected expecting then | 1 |
15,696 | 3,971,874,389 | IssuesEvent | 2016-05-04 13:37:52 | MarlinFirmware/Marlin | https://api.github.com/repos/MarlinFirmware/Marlin | closed | Tools for debugging Marlin? | Discussion: Developers Status: Inactive Support: Documentation Support: Hardware | I was trying to determine why ABL was not working as expected for my Markerfarm i3v printer by using the VisualMicro Debugger with Visual Studio 13 or Atmel Studio 6.2 software (Windows 7 64-bit). I am able to compile the Marlin release build successfully and download to the Atmega2560 board. However, when I try to compile a debug build, the compile fails as follows:
```shell
Compiling debug version of 'Marlin' for 'Arduino Mega w/ ATmega2560 (Mega 2560)'
Build folder: file:///C:/Users/Pete/AppData/Local/V.Micro/Arduino/Builds/Marlin/mega_atmega2560
VM_DBG.h:153:2: error: 'HardwareSerial' does not name a type
VM_DBG.h:154:13: error: 'HardwareSerial' has not been declared
VM_DBG.h:155:13: error: 'HardwareSerial' has not been declared
VM_DBG.h:In member function 'void VisualMicroDebug::setVariable(T&, int, int)'
VM_DBG.h:323:2: error: 'transport' was not declared in this scope
Error compiling
```
I'm guessing that its related to the MarlinSerial implementation of HardwareSerial module.
Has anyone resolved this issue or can suggest a work around?
I just received my Atmel-ICE programmer/debugger hardware and will try that next.
| 1.0 | Tools for debugging Marlin? - I was trying to determine why ABL was not working as expected for my Markerfarm i3v printer by using the VisualMicro Debugger with Visual Studio 13 or Atmel Studio 6.2 software (Windows 7 64-bit). I am able to compile the Marlin release build successfully and download to the Atmega2560 board. However, when I try to compile a debug build, the compile fails as follows:
```shell
Compiling debug version of 'Marlin' for 'Arduino Mega w/ ATmega2560 (Mega 2560)'
Build folder: file:///C:/Users/Pete/AppData/Local/V.Micro/Arduino/Builds/Marlin/mega_atmega2560
VM_DBG.h:153:2: error: 'HardwareSerial' does not name a type
VM_DBG.h:154:13: error: 'HardwareSerial' has not been declared
VM_DBG.h:155:13: error: 'HardwareSerial' has not been declared
VM_DBG.h:In member function 'void VisualMicroDebug::setVariable(T&, int, int)'
VM_DBG.h:323:2: error: 'transport' was not declared in this scope
Error compiling
```
I'm guessing that its related to the MarlinSerial implementation of HardwareSerial module.
Has anyone resolved this issue or can suggest a work around?
I just received my Atmel-ICE programmer/debugger hardware and will try that next.
| non_priority | tools for debugging marlin i was trying to determine why abl was not working as expected for my markerfarm printer by using the visualmicro debugger with visual studio or atmel studio software windows bit i am able to compile the marlin release build successfully and download to the board however when i try to compile a debug build the compile fails as follows shell compiling debug version of marlin for arduino mega w mega build folder file c users pete appdata local v micro arduino builds marlin mega vm dbg h error hardwareserial does not name a type vm dbg h error hardwareserial has not been declared vm dbg h error hardwareserial has not been declared vm dbg h in member function void visualmicrodebug setvariable t int int vm dbg h error transport was not declared in this scope error compiling i m guessing that its related to the marlinserial implementation of hardwareserial module has anyone resolved this issue or can suggest a work around i just received my atmel ice programmer debugger hardware and will try that next | 0 |
1,616 | 4,166,601,320 | IssuesEvent | 2016-06-20 04:51:15 | couchbase/sync_gateway | https://api.github.com/repos/couchbase/sync_gateway | closed | POST to _changes is incompatible with CouchDB's implementation | compatibility couchdb enhancement P1: high release-prep review | So it turns out CouchDB also added support for POST to `_changes`, only they did it differently โ they still require all the parameters to be in the URL except for the filter params. (I don't know if they weren't aware of our [spec](https://github.com/couchbase/sync_gateway/wiki/POST-to-_changes), or if they were but just didn't follow it.)
The result is that if a client sends a request in the CouchDB style, SG will ignore all of the parameters except for the filter params. That means it doesn't see the `since`, `limit`, etc ... fairly bad.
For compatibility purposes, it would be nice if SG could parse the URL as well as the request body.
(This was first filed as couchbase/couchbase-lite-ios#1139.) | True | POST to _changes is incompatible with CouchDB's implementation - So it turns out CouchDB also added support for POST to `_changes`, only they did it differently โ they still require all the parameters to be in the URL except for the filter params. (I don't know if they weren't aware of our [spec](https://github.com/couchbase/sync_gateway/wiki/POST-to-_changes), or if they were but just didn't follow it.)
The result is that if a client sends a request in the CouchDB style, SG will ignore all of the parameters except for the filter params. That means it doesn't see the `since`, `limit`, etc ... fairly bad.
For compatibility purposes, it would be nice if SG could parse the URL as well as the request body.
(This was first filed as couchbase/couchbase-lite-ios#1139.) | non_priority | post to changes is incompatible with couchdb s implementation so it turns out couchdb also added support for post to changes only they did it differently โ they still require all the parameters to be in the url except for the filter params i don t know if they weren t aware of our or if they were but just didn t follow it the result is that if a client sends a request in the couchdb style sg will ignore all of the parameters except for the filter params that means it doesn t see the since limit etc fairly bad for compatibility purposes it would be nice if sg could parse the url as well as the request body this was first filed as couchbase couchbase lite ios | 0 |
135,784 | 12,691,476,467 | IssuesEvent | 2020-06-21 17:14:12 | Alschn/Binary-Turtle-Graphics | https://api.github.com/repos/Alschn/Binary-Turtle-Graphics | opened | Add documentation | documentation | - Add description of the project including explanation of commands and the way the program works
- Translate documentation to English
- Update readme | 1.0 | Add documentation - - Add description of the project including explanation of commands and the way the program works
- Translate documentation to English
- Update readme | non_priority | add documentation add description of the project including explanation of commands and the way the program works translate documentation to english update readme | 0 |
180,156 | 30,452,810,883 | IssuesEvent | 2023-07-16 14:09:33 | sboxgame/issues | https://api.github.com/repos/sboxgame/issues | opened | Consider Making `internal virtual void Var.Write(NetWrite)` and `internal virtual void Var.Read(NetRead)` Protected | api design | ### What it is?
Currently we're able to network [several different types](https://wiki.facepunch.com/sbox/Networked_Types#networkabletypes), but collections are definitely a sore spot. Implementing whatever collection you need by deriving from `Var` would be great, but because `internal virtual void Write(NetWrite)` and `internal virtual void Read(NetRead)` are internal, it can't be done.
### What should it be?
Making `internal virtual void Write(NetWrite)` and `internal virtual void Read(NetRead)` protected would allow `Var` to be derived, which would add a lot of flexibility with networking. With [`NetRead` being replaced with `ByteSteam`](https://github.com/sboxgame/issues/issues/3620#issuecomment-1636667532), I'd assume `Var` is going to be cleaned up as well; if common collections like `HashSet`, `Stack`, `Queue`, etc. aren't added during that, then we should at least have the ability to implement them ourselves. | 1.0 | Consider Making `internal virtual void Var.Write(NetWrite)` and `internal virtual void Var.Read(NetRead)` Protected - ### What it is?
Currently we're able to network [several different types](https://wiki.facepunch.com/sbox/Networked_Types#networkabletypes), but collections are definitely a sore spot. Implementing whatever collection you need by deriving from `Var` would be great, but because `internal virtual void Write(NetWrite)` and `internal virtual void Read(NetRead)` are internal, it can't be done.
### What should it be?
Making `internal virtual void Write(NetWrite)` and `internal virtual void Read(NetRead)` protected would allow `Var` to be derived, which would add a lot of flexibility with networking. With [`NetRead` being replaced with `ByteSteam`](https://github.com/sboxgame/issues/issues/3620#issuecomment-1636667532), I'd assume `Var` is going to be cleaned up as well; if common collections like `HashSet`, `Stack`, `Queue`, etc. aren't added during that, then we should at least have the ability to implement them ourselves. | non_priority | consider making internal virtual void var write netwrite and internal virtual void var read netread protected what it is currently we re able to network but collections are definitely a sore spot implementing whatever collection you need by deriving from var would be great but because internal virtual void write netwrite and internal virtual void read netread are internal it can t be done what should it be making internal virtual void write netwrite and internal virtual void read netread protected would allow var to be derived which would add a lot of flexibility with networking with i d assume var is going to be cleaned up as well if common collections like hashset stack queue etc aren t added during that then we should at least have the ability to implement them ourselves | 0 |
829,044 | 31,852,062,401 | IssuesEvent | 2023-09-15 03:05:48 | NIAEFEUP/tts-be | https://api.github.com/repos/NIAEFEUP/tts-be | closed | Setup SQLite for Analysis tool | medium effort medium priority | Setup and dockerize an SQLite database so it can be used in the future to cache the backend statistics. | 1.0 | Setup SQLite for Analysis tool - Setup and dockerize an SQLite database so it can be used in the future to cache the backend statistics. | priority | setup sqlite for analysis tool setup and dockerize an sqlite database so it can be used in the future to cache the backend statistics | 1 |
135,297 | 5,246,055,345 | IssuesEvent | 2017-02-01 08:04:15 | domaindrivendev/Swashbuckle.AspNetCore | https://api.github.com/repos/domaindrivendev/Swashbuckle.AspNetCore | closed | Support for ASP.NET API Versioning? | high priority | Hi, in of our .Net Core API's I implemented [Microsoft ASP.NET API Versioning](https://github.com/Microsoft/aspnet-api-versioning), but it seems it's not picked up by Swagger.
Code example for a v2 version:
````cs
[ApiVersion("2")]
[Route("api/v{version:apiVersion}/{tenantid}/timeline")]
public class TimelineController : Controller
{
}
````
This results in

with this Swagger json that also contains strange data:
````javascript
{
"swagger": "2.0",
"info": {
"version": "v2",
},
"basePath": "/",
"paths": {
"/api/v{version}/{tenantid}/timeline/default": {
"get": {
"tags": ["Timeline"],
"operationId": "ApiV{versionByTenantidTimelineDefaultGet",
"parameters": [{
"name": "version",
"in": "path",
"required": true,
"type": "string"
}],
}
}
}
}
````
The version is now treated as a parameter. Can I fix this by configuration? Or am I doing something wrong? | 1.0 | Support for ASP.NET API Versioning? - Hi, in of our .Net Core API's I implemented [Microsoft ASP.NET API Versioning](https://github.com/Microsoft/aspnet-api-versioning), but it seems it's not picked up by Swagger.
Code example for a v2 version:
````cs
[ApiVersion("2")]
[Route("api/v{version:apiVersion}/{tenantid}/timeline")]
public class TimelineController : Controller
{
}
````
This results in

with this Swagger json that also contains strange data:
````javascript
{
"swagger": "2.0",
"info": {
"version": "v2",
},
"basePath": "/",
"paths": {
"/api/v{version}/{tenantid}/timeline/default": {
"get": {
"tags": ["Timeline"],
"operationId": "ApiV{versionByTenantidTimelineDefaultGet",
"parameters": [{
"name": "version",
"in": "path",
"required": true,
"type": "string"
}],
}
}
}
}
````
The version is now treated as a parameter. Can I fix this by configuration? Or am I doing something wrong? | priority | support for asp net api versioning hi in of our net core api s i implemented but it seems it s not picked up by swagger code example for a version cs public class timelinecontroller controller this results in with this swagger json that also contains strange data javascript swagger info version basepath paths api v version tenantid timeline default get tags operationid apiv versionbytenantidtimelinedefaultget parameters name version in path required true type string the version is now treated as a parameter can i fix this by configuration or am i doing something wrong | 1 |
126,388 | 17,876,733,458 | IssuesEvent | 2021-09-07 05:35:06 | Srinivasanms16/EmployeeInformation | https://api.github.com/repos/Srinivasanms16/EmployeeInformation | opened | WS-2019-0424 (Medium) detected in elliptic-6.5.2.tgz | security vulnerability | ## WS-2019-0424 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.2.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz</a></p>
<p>Path to dependency file: EmployeeInformation/package.json</p>
<p>Path to vulnerable library: EmployeeInformation/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.900.7.tgz (Root Library)
- webpack-4.41.2.tgz
- node-libs-browser-2.2.1.tgz
- crypto-browserify-3.12.0.tgz
- browserify-sign-4.0.4.tgz
- :x: **elliptic-6.5.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Srinivasanms16/EmployeeInformation/commit/ea3536ba0d959fc4f94bf484768cf26c70ed48cb">ea3536ba0d959fc4f94bf484768cf26c70ed48cb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
all versions of elliptic are vulnerable to Timing Attack through side-channels.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0424 (Medium) detected in elliptic-6.5.2.tgz - ## WS-2019-0424 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.2.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz</a></p>
<p>Path to dependency file: EmployeeInformation/package.json</p>
<p>Path to vulnerable library: EmployeeInformation/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.900.7.tgz (Root Library)
- webpack-4.41.2.tgz
- node-libs-browser-2.2.1.tgz
- crypto-browserify-3.12.0.tgz
- browserify-sign-4.0.4.tgz
- :x: **elliptic-6.5.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Srinivasanms16/EmployeeInformation/commit/ea3536ba0d959fc4f94bf484768cf26c70ed48cb">ea3536ba0d959fc4f94bf484768cf26c70ed48cb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
all versions of elliptic are vulnerable to Timing Attack through side-channels.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file employeeinformation package json path to vulnerable library employeeinformation node modules elliptic package json dependency hierarchy build angular tgz root library webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of elliptic are vulnerable to timing attack through side channels publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact none for more information on scores click a href step up your open source security game with whitesource | 0 |
4,395 | 5,103,846,187 | IssuesEvent | 2017-01-04 22:43:45 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | SslStream.WriteAsync hangs after updating to CoreFx 4.4 | area-System.Net.Security bug | We've recently updated package references in Kestrel to CoreFx 4.4 https://github.com/aspnet/KestrelHttpServer/commit/5b8f7c2b2be5744edac8ce3c5a2dea86ed65b8ee.
After this update, Kestrel can't serve HTTPS anymore. We have a stream that wraps an `SslStream`, and it's hanging when writing here:
https://github.com/aspnet/KestrelHttpServer/blob/5b8f7c2b2be5744edac8ce3c5a2dea86ed65b8ee/src/Microsoft.AspNetCore.Server.Kestrel/Filter/Internal/StreamSocketOutput.cs#L59
Reproing is straightforward: just run the Kestrel sample app (under `samples/SampleApp` in the repo) and hit it with any client, and you will see the hang.
This only happens on .NET Core. Desktop .NET works fine. | True | SslStream.WriteAsync hangs after updating to CoreFx 4.4 - We've recently updated package references in Kestrel to CoreFx 4.4 https://github.com/aspnet/KestrelHttpServer/commit/5b8f7c2b2be5744edac8ce3c5a2dea86ed65b8ee.
After this update, Kestrel can't serve HTTPS anymore. We have a stream that wraps an `SslStream`, and it's hanging when writing here:
https://github.com/aspnet/KestrelHttpServer/blob/5b8f7c2b2be5744edac8ce3c5a2dea86ed65b8ee/src/Microsoft.AspNetCore.Server.Kestrel/Filter/Internal/StreamSocketOutput.cs#L59
Reproing is straightforward: just run the Kestrel sample app (under `samples/SampleApp` in the repo) and hit it with any client, and you will see the hang.
This only happens on .NET Core. Desktop .NET works fine. | non_priority | sslstream writeasync hangs after updating to corefx we ve recently updated package references in kestrel to corefx after this update kestrel can t serve https anymore we have a stream that wraps an sslstream and it s hanging when writing here reproing is straightforward just run the kestrel sample app under samples sampleapp in the repo and hit it with any client and you will see the hang this only happens on net core desktop net works fine | 0 |
328,999 | 10,010,941,158 | IssuesEvent | 2019-07-15 09:19:32 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | TobyFit Resolution: Verify simulation in comparison with Tobyfit/Horace | Component: Direct Inelastic Misc: Roadmap Priority: High | This issue was originally [TRAC 11050](http://trac.mantidproject.org/mantid/ticket/11050)
This is an umbrella issue to track the issues contributing to the comparison of simulation of the TobyFit resolution function in Vates.
This will be made of sub-issues:
- #11056
- #15556
| 1.0 | TobyFit Resolution: Verify simulation in comparison with Tobyfit/Horace - This issue was originally [TRAC 11050](http://trac.mantidproject.org/mantid/ticket/11050)
This is an umbrella issue to track the issues contributing to the comparison of simulation of the TobyFit resolution function in Vates.
This will be made of sub-issues:
- #11056
- #15556
| priority | tobyfit resolution verify simulation in comparison with tobyfit horace this issue was originally this is an umbrella issue to track the issues contributing to the comparison of simulation of the tobyfit resolution function in vates this will be made of sub issues | 1 |
664,399 | 22,268,774,705 | IssuesEvent | 2022-06-10 10:05:27 | mozilla/perfcompare | https://api.github.com/repos/mozilla/perfcompare | closed | export store creating function for tests | priority: low | from Julien:
> I believe the next step would be to export a store creating function instead of the store object itself, and create an empty store for each test, so that tests are always fully isolated. Happy to discuss about it more later! I was thinking that as a first step, if you don't want to change existing tests, you could export both `store` and a `createStore()` function, so that new tests would use the creator, but old tests would still work with the object directly. That's up to you really :-) | 1.0 | export store creating function for tests - from Julien:
> I believe the next step would be to export a store creating function instead of the store object itself, and create an empty store for each test, so that tests are always fully isolated. Happy to discuss about it more later! I was thinking that as a first step, if you don't want to change existing tests, you could export both `store` and a `createStore()` function, so that new tests would use the creator, but old tests would still work with the object directly. That's up to you really :-) | priority | export store creating function for tests from julien i believe the next step would be to export a store creating function instead of the store object itself and create an empty store for each test so that tests are always fully isolated happy to discuss about it more later i was thinking that as a first step if you don t want to change existing tests you could export both store and a createstore function so that new tests would use the creator but old tests would still work with the object directly that s up to you really | 1 |
16,217 | 4,028,806,649 | IssuesEvent | 2016-05-18 08:12:26 | leekelleher/umbraco-ditto | https://api.github.com/repos/leekelleher/umbraco-ditto | reopened | Processors - update documentation | documentation | We need to document how processors work, (and clean up the ValueResolvers / TypeConverters doco). | 1.0 | Processors - update documentation - We need to document how processors work, (and clean up the ValueResolvers / TypeConverters doco). | non_priority | processors update documentation we need to document how processors work and clean up the valueresolvers typeconverters doco | 0 |
103,908 | 22,499,034,078 | IssuesEvent | 2022-06-23 10:05:05 | ably/ably-java | https://api.github.com/repos/ably/ably-java | closed | Prepare the "lib" module configuration for publishing to Maven Central | enhancement code-quality | As a part of the `2.0.0` release we will implement https://github.com/ably/ably-java/issues/728 and as a result, we will have to publish the newly created module to the Maven Central. To do this, we will need to create a publishing configuration for the newly added `lib` module.
โIssue is synchronized with this [Jira Uncategorised](https://ably.atlassian.net/browse/SDK-1860) by [Unito](https://www.unito.io)
| 1.0 | Prepare the "lib" module configuration for publishing to Maven Central - As a part of the `2.0.0` release we will implement https://github.com/ably/ably-java/issues/728 and as a result, we will have to publish the newly created module to the Maven Central. To do this, we will need to create a publishing configuration for the newly added `lib` module.
โIssue is synchronized with this [Jira Uncategorised](https://ably.atlassian.net/browse/SDK-1860) by [Unito](https://www.unito.io)
| non_priority | prepare the lib module configuration for publishing to maven central as a part of the release we will implement and as a result we will have to publish the newly created module to the maven central to do this we will need to create a publishing configuration for the newly added lib module โissue is synchronized with this by | 0 |
607,336 | 18,780,310,120 | IssuesEvent | 2021-11-08 05:17:02 | python/mypy | https://api.github.com/repos/python/mypy | closed | Duplicate Base Class for TypedDict | topic-typed-dict priority-1-normal topic-usability false-positive | Consider the following code:
```py
from mypy_extensions import TypedDict
class MovieBase(TypedDict):
name: str
year: int
class Movie(MovieBase, MovieBase):
based_on: str
```
Notice that the `Movie` class inherits `MovieBase` twice.
mypy gives the following errors:
```
n.py:8: error: Cannot overwrite TypedDict field "name" while merging
n.py:8: error: Cannot overwrite TypedDict field "year" while merging
```
while there's no runtime error.
Note that if `MovieBase` was not a `TypedDict`, there would be a runtime TypeError: `duplicate base class MovieBase`
I think the right thing here would be to report duplicate base class. | 1.0 | Duplicate Base Class for TypedDict - Consider the following code:
```py
from mypy_extensions import TypedDict
class MovieBase(TypedDict):
name: str
year: int
class Movie(MovieBase, MovieBase):
based_on: str
```
Notice that the `Movie` class inherits `MovieBase` twice.
mypy gives the following errors:
```
n.py:8: error: Cannot overwrite TypedDict field "name" while merging
n.py:8: error: Cannot overwrite TypedDict field "year" while merging
```
while there's no runtime error.
Note that if `MovieBase` was not a `TypedDict`, there would be a runtime TypeError: `duplicate base class MovieBase`
I think the right thing here would be to report duplicate base class. | priority | duplicate base class for typeddict consider the following code py from mypy extensions import typeddict class moviebase typeddict name str year int class movie moviebase moviebase based on str notice that the movie class inherits moviebase twice mypy gives the following errors n py error cannot overwrite typeddict field name while merging n py error cannot overwrite typeddict field year while merging while there s no runtime error note that if moviebase was not a typeddict there would be a runtime typeerror duplicate base class moviebase i think the right thing here would be to report duplicate base class | 1 |
558,385 | 16,532,172,329 | IssuesEvent | 2021-05-27 07:35:28 | Find-U-I/Find-iOS | https://api.github.com/repos/Find-U-I/Find-iOS | closed | [feat] ๋งค์นญ ํํฉ ํญ ๊ฐ๋ฐ(ํธ๊ฐ/์ฐ) | 2๏ธโฃ priority: middle ๐ด view | ## ๐ก Issue
๋งค์นญ ํํฉ ํญ ๊ฐ๋ฐ ์์ํ ๊ฒ์
## ๐ todo
- [x] ํธ๊ฐ/์ฐ ์ธ๊ทธ
- [x] ์ฐ๊ฒฐ๋ ์๋
- [x] ๋ณด๋ธ ํธ๊ฐ
- [x] ๋ฐ์ ํธ๊ฐ
- [x] ๋๋ฅผ ์ฐํ ์ฌ๋
- [x] ๋ด๊ฐ ์ฐํ ์ฌ๋
- [x] ์ธ๊ทธ ์ ํ ๋ถ๋๋ฝ๊ฒ..
| 1.0 | [feat] ๋งค์นญ ํํฉ ํญ ๊ฐ๋ฐ(ํธ๊ฐ/์ฐ) - ## ๐ก Issue
๋งค์นญ ํํฉ ํญ ๊ฐ๋ฐ ์์ํ ๊ฒ์
## ๐ todo
- [x] ํธ๊ฐ/์ฐ ์ธ๊ทธ
- [x] ์ฐ๊ฒฐ๋ ์๋
- [x] ๋ณด๋ธ ํธ๊ฐ
- [x] ๋ฐ์ ํธ๊ฐ
- [x] ๋๋ฅผ ์ฐํ ์ฌ๋
- [x] ๋ด๊ฐ ์ฐํ ์ฌ๋
- [x] ์ธ๊ทธ ์ ํ ๋ถ๋๋ฝ๊ฒ..
| priority | ๋งค์นญ ํํฉ ํญ ๊ฐ๋ฐ ํธ๊ฐ ์ฐ ๐ก issue ๋งค์นญ ํํฉ ํญ ๊ฐ๋ฐ ์์ํ ๊ฒ์ ๐ todo ํธ๊ฐ ์ฐ ์ธ๊ทธ ์ฐ๊ฒฐ๋ ์๋ ๋ณด๋ธ ํธ๊ฐ ๋ฐ์ ํธ๊ฐ ๋๋ฅผ ์ฐํ ์ฌ๋ ๋ด๊ฐ ์ฐํ ์ฌ๋ ์ธ๊ทธ ์ ํ ๋ถ๋๋ฝ๊ฒ | 1 |
196,571 | 22,442,147,930 | IssuesEvent | 2022-06-21 02:34:55 | valdisiljuconoks/episerverless | https://api.github.com/repos/valdisiljuconoks/episerverless | closed | WS-2019-0332 (High) detected in handlebars-1.3.0.tgz - autoclosed | security vulnerability | ## WS-2019-0332 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p>
<p>Path to dependency file: episerverless/src/Web/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json,episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- istanbul-0.2.16.tgz
- :x: **handlebars-1.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331.
<p>Publish Date: 2019-11-17
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7>WS-2019-0332</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p>
<p>Release Date: 2019-11-17</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0332 (High) detected in handlebars-1.3.0.tgz - autoclosed - ## WS-2019-0332 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p>
<p>Path to dependency file: episerverless/src/Web/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json,episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- istanbul-0.2.16.tgz
- :x: **handlebars-1.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331.
<p>Publish Date: 2019-11-17
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7>WS-2019-0332</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p>
<p>Release Date: 2019-11-17</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws high detected in handlebars tgz autoclosed ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file episerverless src web modules protected shell shell clientresources lib xstyle package json path to vulnerable library episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle node modules handlebars package json episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle node modules handlebars package json dependency hierarchy intern geezer tgz root library istanbul tgz x handlebars tgz vulnerable library vulnerability details arbitrary code execution vulnerability found in handlebars before lookup helper fails to validate templates attack may submit templates that execute arbitrary javascript in the system it is due to an incomplete fix for a ws publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource | 0 |
41,341 | 2,868,999,142 | IssuesEvent | 2015-06-05 22:28:46 | dart-lang/args | https://api.github.com/repos/dart-lang/args | opened | zsh/bash autocompletion generator from ArgParser | enhancement Priority-Low | <a href="https://github.com/amouravski"><img src="https://avatars.githubusercontent.com/u/264967?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [amouravski](https://github.com/amouravski)**
_Originally opened as dart-lang/sdk#8389_
----
zsh and bash both have a shell autocomplete capability, so that you can quickly complete flags, arguments, etc.
It'd be amazing if we could ingest an ArgParser and spit out zsh/bash completion files so that any dart script can autocomplete in this way. | 1.0 | zsh/bash autocompletion generator from ArgParser - <a href="https://github.com/amouravski"><img src="https://avatars.githubusercontent.com/u/264967?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [amouravski](https://github.com/amouravski)**
_Originally opened as dart-lang/sdk#8389_
----
zsh and bash both have a shell autocomplete capability, so that you can quickly complete flags, arguments, etc.
It'd be amazing if we could ingest an ArgParser and spit out zsh/bash completion files so that any dart script can autocomplete in this way. | priority | zsh bash autocompletion generator from argparser issue by originally opened as dart lang sdk zsh and bash both have a shell autocomplete capability so that you can quickly complete flags arguments etc it d be amazing if we could ingest an argparser and spit out zsh bash completion files so that any dart script can autocomplete in this way | 1 |
919 | 3,378,314,489 | IssuesEvent | 2015-11-25 10:11:56 | Wikitalia/edgesense | https://api.github.com/repos/Wikitalia/edgesense | closed | Remove loops from graph visualization and edge count | enhancement front-end processing | It is not clear that loops have meaning in a social network of comments. Removing them improves the clarity of the graph. | 1.0 | Remove loops from graph visualization and edge count - It is not clear that loops have meaning in a social network of comments. Removing them improves the clarity of the graph. | non_priority | remove loops from graph visualization and edge count it is not clear that loops have meaning in a social network of comments removing them improves the clarity of the graph | 0 |
47,548 | 13,056,236,650 | IssuesEvent | 2020-07-30 04:04:53 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | blas/lapack need check_language() and enable_language() (Trac #724) | Migrated from Trac cmake defect | ... especially for SYSTEM_TOOLS support
Migrated from https://code.icecube.wisc.edu/ticket/724
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:22",
"description": "... especially for SYSTEM_TOOLS support",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"_ts": "1550067082284240",
"component": "cmake",
"summary": "blas/lapack need check_language() and enable_language()",
"priority": "normal",
"keywords": "",
"time": "2014-04-14T19:30:53",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | blas/lapack need check_language() and enable_language() (Trac #724) - ... especially for SYSTEM_TOOLS support
Migrated from https://code.icecube.wisc.edu/ticket/724
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:22",
"description": "... especially for SYSTEM_TOOLS support",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"_ts": "1550067082284240",
"component": "cmake",
"summary": "blas/lapack need check_language() and enable_language()",
"priority": "normal",
"keywords": "",
"time": "2014-04-14T19:30:53",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| non_priority | blas lapack need check language and enable language trac especially for system tools support migrated from json status closed changetime description especially for system tools support reporter nega cc resolution worksforme ts component cmake summary blas lapack need check language and enable language priority normal keywords time milestone owner nega type defect | 0 |
65,469 | 27,110,866,646 | IssuesEvent | 2023-02-15 15:13:50 | hashicorp/terraform-provider-azurerm | https://api.github.com/repos/hashicorp/terraform-provider-azurerm | closed | azurerm_billing_account_cost_management_export Status=404 Code="MissingSubscription" | bug service/billing | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.6
### AzureRM Provider Version
3.41.0
### Affected Resource(s)/Data Source(s)
azurerm_billing_account_cost_management_export
### Terraform Configuration Files
```hcl
resource "azurerm_storage_account" "storageaccount" {
provider = azurerm.management
name = "storageaccount"
resource_group_name = "CostManagement"
location = "West US 2"
account_tier = "Standard"
account_replication_type = "LRS"
allow_nested_items_to_be_public = false
}
resource "azurerm_storage_container" "actual-billingaccount" {
provider = azurerm.management
name = "actual-billingaccount"
storage_account_name = azurerm_storage_account.storageaccount.name
container_access_type = "private"
}
resource "azurerm_billing_account_cost_management_export" "billing_account_export" {
name = "actual-billingaccount"
billing_account_id = "<BillingAccountId>/billingProfiles/<BillingProfileId>"
recurrence_type = "Daily"
recurrence_period_start_date = formatdate("YYYY-MM-DD'T'00:00:00Z", timestamp())
recurrence_period_end_date = "2050-01-01T00:00:00Z"
export_data_storage_location {
container_id = azurerm_storage_container.actual-billingaccount.resource_manager_id
root_folder_path = "/billingaccount-actual"
}
export_data_options {
type = "ActualCost"
time_frame = "WeekToDate"
}
}
```
### Debug Output/Panic Output
```shell
https://gist.github.com/danabrash/56f702a06d6339ccc21aea277334f298
```
### Expected Behaviour
The export job should create. There is no documented setting for context beyond the billing account ID.
Since we're MCA (not EA) I tried both just the billing account ID and the billing accountID/billingprofileID, same results
The prompt for subscription or Tenant is out of context
### Actual Behaviour
Status=404 Code="MissingSubscription" Message="The request did not have a subscription or a valid tenant level resource provider."
### Steps to Reproduce
run the sample code provided. substitute actual values for account tokens
### Important Factoids
_No response_
### References
_No response_ | 1.0 | azurerm_billing_account_cost_management_export Status=404 Code="MissingSubscription" - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.6
### AzureRM Provider Version
3.41.0
### Affected Resource(s)/Data Source(s)
azurerm_billing_account_cost_management_export
### Terraform Configuration Files
```hcl
resource "azurerm_storage_account" "storageaccount" {
provider = azurerm.management
name = "storageaccount"
resource_group_name = "CostManagement"
location = "West US 2"
account_tier = "Standard"
account_replication_type = "LRS"
allow_nested_items_to_be_public = false
}
resource "azurerm_storage_container" "actual-billingaccount" {
provider = azurerm.management
name = "actual-billingaccount"
storage_account_name = azurerm_storage_account.storageaccount.name
container_access_type = "private"
}
resource "azurerm_billing_account_cost_management_export" "billing_account_export" {
name = "actual-billingaccount"
billing_account_id = "<BillingAccountId>/billingProfiles/<BillingProfileId>"
recurrence_type = "Daily"
recurrence_period_start_date = formatdate("YYYY-MM-DD'T'00:00:00Z", timestamp())
recurrence_period_end_date = "2050-01-01T00:00:00Z"
export_data_storage_location {
container_id = azurerm_storage_container.actual-billingaccount.resource_manager_id
root_folder_path = "/billingaccount-actual"
}
export_data_options {
type = "ActualCost"
time_frame = "WeekToDate"
}
}
```
### Debug Output/Panic Output
```shell
https://gist.github.com/danabrash/56f702a06d6339ccc21aea277334f298
```
### Expected Behaviour
The export job should create. There is no documented setting for context beyond the billing account ID.
Since we're MCA (not EA) I tried both just the billing account ID and the billing accountID/billingprofileID, same results
The prompt for subscription or Tenant is out of context
### Actual Behaviour
Status=404 Code="MissingSubscription" Message="The request did not have a subscription or a valid tenant level resource provider."
### Steps to Reproduce
run the sample code provided. substitute actual values for account tokens
### Important Factoids
_No response_
### References
_No response_ | non_priority | azurerm billing account cost management export status code missingsubscription is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm billing account cost management export terraform configuration files hcl resource azurerm storage account storageaccount provider azurerm management name storageaccount resource group name costmanagement location west us account tier standard account replication type lrs allow nested items to be public false resource azurerm storage container actual billingaccount provider azurerm management name actual billingaccount storage account name azurerm storage account storageaccount name container access type private resource azurerm billing account cost management export billing account export name actual billingaccount billing account id billingprofiles recurrence type daily recurrence period start date formatdate yyyy mm dd t timestamp recurrence period end date export data storage location container id azurerm storage container actual billingaccount resource manager id root folder path billingaccount actual export data options type actualcost time frame weektodate debug output panic output shell expected behaviour the export job should create there is no documented setting for context beyond the billing account id since we re mca not ea i tried both just the billing account id and the billing accountid billingprofileid same results the prompt for subscription or tenant is out of context actual behaviour status code missingsubscription message the request did not have a subscription or a valid tenant level resource provider steps to reproduce run the sample code provided substitute actual values for account tokens important factoids no response references no response | 0 |
830,971 | 32,032,999,754 | IssuesEvent | 2023-09-22 13:34:06 | ApplETS/Notre-Dame | https://api.github.com/repos/ApplETS/Notre-Dame | closed | Reorderable links in Quicklinks | enhancement feature: รTS priority: medium ready to develop | **Is your feature request related to a problem? Please describe.**
It would be nice to have a reorderable quicklinks page so all links can be reorder, remove or added. For this to work, we would need to maintain 1 list of all links (never changes) and one list of currently displayed links (part of the complete list). An index of each link must be saved to be able to track where the item is located.
We would also need to save theses infos in localstorage.
**Acceptance criteria**
- [ ] On page open, order the list based on the order saved;
- [ ] On long press of a quicklink, activate the reactivate feature and show small 'x' to delete a link;
- [ ] Add a way to create a link that is disappear (can't create 2 identical links too);
- [ ] On a press anywhere else without links, we should save once again the quicklinks with the order in localstorage;
- [ ] Make sure to have an initializer to insert default layout when none is found;
- [ ] Tests ๐งช
**Describe alternatives you've considered**
N/A
**Additional context**
This package could be easily used ๐ https://pub.dev/packages/reorderable_grid_view
Also, it's to note that with the new horairรTS link, this feature is more urgent now since we don't necessarily want to schedule genrator at the same time. | 1.0 | Reorderable links in Quicklinks - **Is your feature request related to a problem? Please describe.**
It would be nice to have a reorderable quicklinks page so all links can be reorder, remove or added. For this to work, we would need to maintain 1 list of all links (never changes) and one list of currently displayed links (part of the complete list). An index of each link must be saved to be able to track where the item is located.
We would also need to save theses infos in localstorage.
**Acceptance criteria**
- [ ] On page open, order the list based on the order saved;
- [ ] On long press of a quicklink, activate the reactivate feature and show small 'x' to delete a link;
- [ ] Add a way to create a link that is disappear (can't create 2 identical links too);
- [ ] On a press anywhere else without links, we should save once again the quicklinks with the order in localstorage;
- [ ] Make sure to have an initializer to insert default layout when none is found;
- [ ] Tests ๐งช
**Describe alternatives you've considered**
N/A
**Additional context**
This package could be easily used ๐ https://pub.dev/packages/reorderable_grid_view
Also, it's to note that with the new horairรTS link, this feature is more urgent now since we don't necessarily want to schedule genrator at the same time. | priority | reorderable links in quicklinks is your feature request related to a problem please describe it would be nice to have a reorderable quicklinks page so all links can be reorder remove or added for this to work we would need to maintain list of all links never changes and one list of currently displayed links part of the complete list an index of each link must be saved to be able to track where the item is located we would also need to save theses infos in localstorage acceptance criteria on page open order the list based on the order saved on long press of a quicklink activate the reactivate feature and show small x to delete a link add a way to create a link that is disappear can t create identical links too on a press anywhere else without links we should save once again the quicklinks with the order in localstorage make sure to have an initializer to insert default layout when none is found tests ๐งช describe alternatives you ve considered n a additional context this package could be easily used ๐ also it s to note that with the new horairรฉts link this feature is more urgent now since we don t necessarily want to schedule genrator at the same time | 1 |
266,024 | 23,217,672,580 | IssuesEvent | 2022-08-02 15:18:11 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | [Flaky Test] Storage Function Bindings: CanExtendVisibilityTimeoutMultipleTimes | Storage Client Functions test-reliability | **Error message**
Expected: True
But was: False
**Stack trace**
at Microsoft.Azure.WebJobs.Extensions.Storage.Queues.Tests.UpdateQueueMessageVisibilityCommandTests.CanExtendVisibilityTimeoutMultipleTimes() in /mnt/vss/_work/1/s/sdk/storage/Microsoft.Azure.WebJobs.Extensions.Storage.Queues/tests/UpdateQueueMessageVisibilityCommandTests.cs:line 51
at NUnit.Framework.Internal.TaskAwaitAdapter.GenericAdapter`1.BlockUntilCompleted()
at NUnit.Framework.Internal.MessagePumpStrategy.NoMessagePumpStrategy.WaitForCompletion(AwaitAdapter awaiter)
at NUnit.Framework.Internal.AsyncToSyncAdapter.Await(Func`1 invoke)
at NUnit.Framework.Internal.Commands.TestMethodCommand.RunTestMethod(TestExecutionContext context)
at NUnit.Framework.Internal.Commands.TestMethodCommand.Execute(TestExecutionContext context)
at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.<>c__DisplayClass1_0.<Execute>b__0()
at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.RunTestMethodInThreadAbortSafeZone(TestExecutionContext context, Action action)
https://dev.azure.com/azure-sdk/public/_build/results?buildId=1075739&view=ms.vss-test-web.build-test-results-tab&runId=23316068&resultId=123571&paneView=debug | 1.0 | [Flaky Test] Storage Function Bindings: CanExtendVisibilityTimeoutMultipleTimes - **Error message**
Expected: True
But was: False
**Stack trace**
at Microsoft.Azure.WebJobs.Extensions.Storage.Queues.Tests.UpdateQueueMessageVisibilityCommandTests.CanExtendVisibilityTimeoutMultipleTimes() in /mnt/vss/_work/1/s/sdk/storage/Microsoft.Azure.WebJobs.Extensions.Storage.Queues/tests/UpdateQueueMessageVisibilityCommandTests.cs:line 51
at NUnit.Framework.Internal.TaskAwaitAdapter.GenericAdapter`1.BlockUntilCompleted()
at NUnit.Framework.Internal.MessagePumpStrategy.NoMessagePumpStrategy.WaitForCompletion(AwaitAdapter awaiter)
at NUnit.Framework.Internal.AsyncToSyncAdapter.Await(Func`1 invoke)
at NUnit.Framework.Internal.Commands.TestMethodCommand.RunTestMethod(TestExecutionContext context)
at NUnit.Framework.Internal.Commands.TestMethodCommand.Execute(TestExecutionContext context)
at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.<>c__DisplayClass1_0.<Execute>b__0()
at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.RunTestMethodInThreadAbortSafeZone(TestExecutionContext context, Action action)
https://dev.azure.com/azure-sdk/public/_build/results?buildId=1075739&view=ms.vss-test-web.build-test-results-tab&runId=23316068&resultId=123571&paneView=debug | non_priority | storage function bindings canextendvisibilitytimeoutmultipletimes error message expected true but was false stack trace at microsoft azure webjobs extensions storage queues tests updatequeuemessagevisibilitycommandtests canextendvisibilitytimeoutmultipletimes in mnt vss work s sdk storage microsoft azure webjobs extensions storage queues tests updatequeuemessagevisibilitycommandtests cs line at nunit framework internal taskawaitadapter genericadapter blockuntilcompleted at nunit framework internal messagepumpstrategy nomessagepumpstrategy waitforcompletion awaitadapter awaiter at nunit framework internal asynctosyncadapter await func invoke at nunit framework internal commands testmethodcommand runtestmethod testexecutioncontext context at nunit framework internal commands testmethodcommand execute testexecutioncontext context at nunit framework internal commands beforeandaftertestcommand c b at nunit framework internal commands beforeandaftertestcommand runtestmethodinthreadabortsafezone testexecutioncontext context action action | 0 |
766,762 | 26,897,727,603 | IssuesEvent | 2023-02-06 13:39:19 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | pixiv.net - see bug description | browser-firefox priority-important engine-gecko | <!-- @browser: Firefox 109.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/109.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/117892 -->
**URL**: https://pixiv.net
**Browser / Version**: Firefox 109.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Something else
**Description**: Copying images to Discord does not copy the image
**Steps to Reproduce**:
Copying an image from Pixiv (pixiv.net) to Discord (discord.com), the file is much smaller than it should be and does not actually show an image, instead it seems more like an empty file. This was not an issue in the past, and is not the case when using Microsoft Edge.
The image will paste correctly into Paint.NET (getpaint.net), however.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | 1.0 | pixiv.net - see bug description - <!-- @browser: Firefox 109.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/109.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/117892 -->
**URL**: https://pixiv.net
**Browser / Version**: Firefox 109.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Something else
**Description**: Copying images to Discord does not copy the image
**Steps to Reproduce**:
Copying an image from Pixiv (pixiv.net) to Discord (discord.com), the file is much smaller than it should be and does not actually show an image, instead it seems more like an empty file. This was not an issue in the past, and is not the case when using Microsoft Edge.
The image will paste correctly into Paint.NET (getpaint.net), however.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | priority | pixiv net see bug description url browser version firefox operating system windows tested another browser yes edge problem type something else description copying images to discord does not copy the image steps to reproduce copying an image from pixiv pixiv net to discord discord com the file is much smaller than it should be and does not actually show an image instead it seems more like an empty file this was not an issue in the past and is not the case when using microsoft edge the image will paste correctly into paint net getpaint net however browser configuration none from with โค๏ธ | 1 |
644,578 | 20,981,487,834 | IssuesEvent | 2022-03-28 20:25:16 | GoogleChrome/lighthouse | https://api.github.com/repos/GoogleChrome/lighthouse | reopened | Lighthouse performance html report loads forever | bug needs-priority | ### FAQ
- [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/master/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/master/docs/throttling.md).
- [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead).
### URL
https://lightlab-stas.vercel.app/
### What happened?
When I run a performance audit, it runs normally and gets stuck at "Almost there! Lighthouse is now generating your report".

### What did you expect?
The performance should show up at the end of the audit
### What have you tried?
The other audits work normally and show the report as expected.
The issue is present on both mobile and desktop.
The issue is only present on my website https://lightlab-stas.vercel.app/ it runs normally on other websites
### How were you running Lighthouse?
Chrome DevTools
### Lighthouse Version
9.2.0
### Chrome Version
99.0.4844.84 (Official Build) (64-bit)
### Node Version
_No response_
### OS
Windows 10
### Relevant log output
```shell
I didn't find lighthouse's log
```
| 1.0 | Lighthouse performance html report loads forever - ### FAQ
- [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/master/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/master/docs/throttling.md).
- [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead).
### URL
https://lightlab-stas.vercel.app/
### What happened?
When I run a performance audit, it runs normally and gets stuck at "Almost there! Lighthouse is now generating your report".

### What did you expect?
The performance should show up at the end of the audit
### What have you tried?
The other audits work normally and show the report as expected.
The issue is present on both mobile and desktop.
The issue is only present on my website https://lightlab-stas.vercel.app/ it runs normally on other websites
### How were you running Lighthouse?
Chrome DevTools
### Lighthouse Version
9.2.0
### Chrome Version
99.0.4844.84 (Official Build) (64-bit)
### Node Version
_No response_
### OS
Windows 10
### Relevant log output
```shell
I didn't find lighthouse's log
```
| priority | lighthouse performance html report loads forever faq yes my issue is not about or yes my issue is not about a specific accessibility audit file with instead url what happened when i run a performance audit it runs normally and gets stuck at almost there lighthouse is now generating your report what did you expect the performance should show up at the end of the audit what have you tried the other audits work normally and show the report as expected the issue is present on both mobile and desktop the issue is only present on my website it runs normally on other websites how were you running lighthouse chrome devtools lighthouse version chrome version official build bit node version no response os windows relevant log output shell i didn t find lighthouse s log | 1 |
737,172 | 25,504,416,975 | IssuesEvent | 2022-11-28 08:15:04 | BVPyro/front | https://api.github.com/repos/BVPyro/front | closed | create bvpk.org/sicher-und-bunt | enhancement high priority | - create bvpk.org/sicher-und-bunt
- link to a cms-accessible page
- might look similar to bvpk.org/mitgliederversammlung
- maybe the space for text etc. could be a little wider...? | 1.0 | create bvpk.org/sicher-und-bunt - - create bvpk.org/sicher-und-bunt
- link to a cms-accessible page
- might look similar to bvpk.org/mitgliederversammlung
- maybe the space for text etc. could be a little wider...? | priority | create bvpk org sicher und bunt create bvpk org sicher und bunt link to a cms accessible page might look similar to bvpk org mitgliederversammlung maybe the space for text etc could be a little wider | 1 |
12,587 | 2,711,585,104 | IssuesEvent | 2015-04-09 07:40:43 | Chris-Ther/deep-learning-faces | https://api.github.com/repos/Chris-Ther/deep-learning-faces | closed | compiling error on make mexf=para | auto-migrated Priority-Medium Type-Defect | ```
I encountered an error on compiling step.
I'm using Cuda 5.5, on ubuntu 12.04 64 bit
I'd followed the steps written in README,
Once I type "make mexf="./deep_nn/mexcuConvNNoo.mex
./deep_nn/mexcuConvNNooFF.mex",
it returns the error:
make: /pkgs/matlab-80/bin/mexext: Command not found
make: *** No rule to make target 'deep_nn/mexcuConvNNoo.', needed by 'all'.
Stop.
I've tried to figure it myself, but it seems to be difficult to me.
Sorry for that I have lack of knowledge on this, I'll doing on my best.
Sincerely
Tae-Ho Kim
```
Original issue reported on code.google.com by `ktho...@gmail.com` on 23 Nov 2013 at 11:59 | 1.0 | compiling error on make mexf=para - ```
I encountered an error on compiling step.
I'm using Cuda 5.5, on ubuntu 12.04 64 bit
I'd followed the steps written in README,
Once I type "make mexf="./deep_nn/mexcuConvNNoo.mex
./deep_nn/mexcuConvNNooFF.mex",
it returns the error:
make: /pkgs/matlab-80/bin/mexext: Command not found
make: *** No rule to make target 'deep_nn/mexcuConvNNoo.', needed by 'all'.
Stop.
I've tried to figure it myself, but it seems to be difficult to me.
Sorry for that I have lack of knowledge on this, I'll doing on my best.
Sincerely
Tae-Ho Kim
```
Original issue reported on code.google.com by `ktho...@gmail.com` on 23 Nov 2013 at 11:59 | non_priority | compiling error on make mexf para i encountered an error on compiling step i m using cuda on ubuntu bit i d followed the steps written in readme once i type make mexf deep nn mexcuconvnnoo mex deep nn mexcuconvnnooff mex it returns the error make pkgs matlab bin mexext command not found make no rule to make target deep nn mexcuconvnnoo needed by all stop i ve tried to figure it myself but it seems to be difficult to me sorry for that i have lack of knowledge on this i ll doing on my best sincerely tae ho kim original issue reported on code google com by ktho gmail com on nov at | 0 |
110,214 | 4,423,423,952 | IssuesEvent | 2016-08-16 08:32:14 | PowerlineApp/powerline-mobile | https://api.github.com/repos/PowerlineApp/powerline-mobile | closed | Facebook Login / API Change | enhancement P1 - High Priority | Per the e-mail I forwarded, the Facebook change requires our action by August 6th. We have time, but we may want to address this as part of the fix for #66 | 1.0 | Facebook Login / API Change - Per the e-mail I forwarded, the Facebook change requires our action by August 6th. We have time, but we may want to address this as part of the fix for #66 | priority | facebook login api change per the e mail i forwarded the facebook change requires our action by august we have time but we may want to address this as part of the fix for | 1 |
379,554 | 11,223,223,777 | IssuesEvent | 2020-01-07 22:06:32 | SETI/pds-opus | https://api.github.com/repos/SETI/pds-opus | closed | "obs_instrument_gossi/frame_duration" has units but not form_type RANGE% | A-Bug B-Import Pipeline Effort 3 Easy Priority 4 Useful | Due to the way GOSSI is implemented in the import pipeline, we always get the warning:
`"obs_instrument_gossi/frame_duration" has units but not form_type RANGE%`
We should find a way to get rid of this warning.
| 1.0 | "obs_instrument_gossi/frame_duration" has units but not form_type RANGE% - Due to the way GOSSI is implemented in the import pipeline, we always get the warning:
`"obs_instrument_gossi/frame_duration" has units but not form_type RANGE%`
We should find a way to get rid of this warning.
| priority | obs instrument gossi frame duration has units but not form type range due to the way gossi is implemented in the import pipeline we always get the warning obs instrument gossi frame duration has units but not form type range we should find a way to get rid of this warning | 1 |
16,624 | 3,547,719,906 | IssuesEvent | 2016-01-20 11:02:28 | I2PC/scipion | https://api.github.com/repos/I2PC/scipion | closed | Test failing for CL2D | bug test | The following test fails:
scipion test tests.em.protocols.test_protocols_xmipp_2d.TestXmippCL2D | 1.0 | Test failing for CL2D - The following test fails:
scipion test tests.em.protocols.test_protocols_xmipp_2d.TestXmippCL2D | non_priority | test failing for the following test fails scipion test tests em protocols test protocols xmipp | 0 |
676,658 | 23,133,095,154 | IssuesEvent | 2022-07-28 12:13:52 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Bookmark icon not updated when question is converted to model and vice-versa | Type:Bug Priority:P3 Organization/ .Frontend Querying/Models | **To Reproduce**
1. Bookmark a question
2. Click on "โฆ" at the top right
3. Click on "Turn into model"
4. Confirm by clicking on "Turn this into model"
Icon on the bookmarks list should change from question to model.
5. Click on "โฆ" at the top right
6. Click on "Turn back into saved question".
Icon on the bookmarks list should change from model to question. | 1.0 | Bookmark icon not updated when question is converted to model and vice-versa - **To Reproduce**
1. Bookmark a question
2. Click on "โฆ" at the top right
3. Click on "Turn into model"
4. Confirm by clicking on "Turn this into model"
Icon on the bookmarks list should change from question to model.
5. Click on "โฆ" at the top right
6. Click on "Turn back into saved question".
Icon on the bookmarks list should change from model to question. | priority | bookmark icon not updated when question is converted to model and vice versa to reproduce bookmark a question click on โฆ at the top right click on turn into model confirm by clicking on turn this into model icon on the bookmarks list should change from question to model click on โฆ at the top right click on turn back into saved question icon on the bookmarks list should change from model to question | 1 |
472,505 | 13,625,980,944 | IssuesEvent | 2020-09-24 10:18:03 | carbon-design-system/carbon-addons-iot-react | https://api.github.com/repos/carbon-design-system/carbon-addons-iot-react | opened | [WizardModal] Add onBack callback | status: needs priority :inbox_tray: status: needs triage :mag: type: enhancement :bulb: | <!--
Use this template if you want to request a new feature, or a change to an
existing feature.
If you'd like to request an entirely new component, please use the component request template instead.
If you are reporting a bug or problem, please use the bug template instead.
-->
### What package is this for?
- [x] React
- [ ] Angular
### Summary
Add a callback prop `onBack` that notifies the parent component when the user goes to a previous step.
**Additional context**
Currently, the wizard modal can only go "forward" by verifying steps. When a user goes back and content needs to be adjusted for that in the parent component, a callback is necessary to notify the parent which previous step has been selected.
| 1.0 | [WizardModal] Add onBack callback - <!--
Use this template if you want to request a new feature, or a change to an
existing feature.
If you'd like to request an entirely new component, please use the component request template instead.
If you are reporting a bug or problem, please use the bug template instead.
-->
### What package is this for?
- [x] React
- [ ] Angular
### Summary
Add a callback prop `onBack` that notifies the parent component when the user goes to a previous step.
**Additional context**
Currently, the wizard modal can only go "forward" by verifying steps. When a user goes back and content needs to be adjusted for that in the parent component, a callback is necessary to notify the parent which previous step has been selected.
| priority | add onback callback use this template if you want to request a new feature or a change to an existing feature if you d like to request an entirely new component please use the component request template instead if you are reporting a bug or problem please use the bug template instead what package is this for react angular summary add a callback prop onback that notifies the parent component when the user goes to a previous step additional context currently the wizard modal can only go forward by verifying steps when a user goes back and content needs to be adjusted for that in the parent component a callback is necessary to notify the parent which previous step has been selected | 1 |
30,534 | 24,899,280,043 | IssuesEvent | 2022-10-28 19:01:03 | GaloisInc/cclyzerpp | https://api.github.com/repos/GaloisInc/cclyzerpp | closed | ci: Compile with `-DCMAKE_BUILD_TYPE=Release` for releases | infrastructure | Releases should be built with optimization. | 1.0 | ci: Compile with `-DCMAKE_BUILD_TYPE=Release` for releases - Releases should be built with optimization. | non_priority | ci compile with dcmake build type release for releases releases should be built with optimization | 0 |
208,861 | 16,165,226,455 | IssuesEvent | 2021-05-01 10:52:27 | argoproj-labs/argocd-operator | https://api.github.com/repos/argoproj-labs/argocd-operator | closed | Document Kustomize Deployment | documentation | We added kustomize to the options for deploying this operator but need to go back and update the documentation to reflect that as well. This is a placeholder to remind me to get that updated. | 1.0 | Document Kustomize Deployment - We added kustomize to the options for deploying this operator but need to go back and update the documentation to reflect that as well. This is a placeholder to remind me to get that updated. | non_priority | document kustomize deployment we added kustomize to the options for deploying this operator but need to go back and update the documentation to reflect that as well this is a placeholder to remind me to get that updated | 0 |
16,169 | 6,122,542,773 | IssuesEvent | 2017-06-23 00:13:41 | oracc/oracc | https://api.github.com/repos/oracc/oracc | closed | Catalogue pager searches fail in P-driven and Score-driven projects | bug build pager search | Here is the original report by Jamie:
I have been trying to get RIBo ready for public viewing. I spent a bit of time last week drafting a page on searching the corpus using the catalogue. See [http://oracc.museum.upenn.edu/ribo/babylon7/usingtheribocorpus/index.html](url)
The catalogue functionality works pretty well on the various projects with Q-number composites: ribo/babylon7, rinap, rinap/rinap1, rinap/rinap3, rinap/rinap4.
For example, object_type:prism yield results.
However, I have yet to perform a success catalogue search in various scores projects and sources projects.
These are:
[http://oracc.museum.upenn.edu/rinap/scores/pager](url)
[http://oracc.museum.upenn.edu/rinap/sources/pager](url)
[http://oracc.museum.upenn.edu/ribo/sources/pager](url)
[http://oracc.museum.upenn.edu/ribo/bab7scores/pager](url)
All have local XML catalogues in their 00cat folders. For example, the ribo/bab7scores catalogue data was extracted from the ribo/babylon7 catalogue, so the search functionality should work for both projects.
For example, in ribo/bab7scores enter ruler:nabopolassar or object:cylinder into the search box. From the Page View, it is clear that these should yield results, but they do not. Either this is a bug or I do not have the correct data in the config.xml files. I attach the config.xml for ribo/bab7scores and ribo/sources. Please advise whether the config need the correct lines or whether this is a bug.
| 1.0 | Catalogue pager searches fail in P-driven and Score-driven projects - Here is the original report by Jamie:
I have been trying to get RIBo ready for public viewing. I spent a bit of time last week drafting a page on searching the corpus using the catalogue. See [http://oracc.museum.upenn.edu/ribo/babylon7/usingtheribocorpus/index.html](url)
The catalogue functionality works pretty well on the various projects with Q-number composites: ribo/babylon7, rinap, rinap/rinap1, rinap/rinap3, rinap/rinap4.
For example, object_type:prism yield results.
However, I have yet to perform a success catalogue search in various scores projects and sources projects.
These are:
[http://oracc.museum.upenn.edu/rinap/scores/pager](url)
[http://oracc.museum.upenn.edu/rinap/sources/pager](url)
[http://oracc.museum.upenn.edu/ribo/sources/pager](url)
[http://oracc.museum.upenn.edu/ribo/bab7scores/pager](url)
All have local XML catalogues in their 00cat folders. For example, the ribo/bab7scores catalogue data was extracted from the ribo/babylon7 catalogue, so the search functionality should work for both projects.
For example, in ribo/bab7scores enter ruler:nabopolassar or object:cylinder into the search box. From the Page View, it is clear that these should yield results, but they do not. Either this is a bug or I do not have the correct data in the config.xml files. I attach the config.xml for ribo/bab7scores and ribo/sources. Please advise whether the config need the correct lines or whether this is a bug.
| non_priority | catalogue pager searches fail in p driven and score driven projects here is the original report by jamie i have been trying to get ribo ready for public viewing i spent a bit of time last week drafting a page on searching the corpus using the catalogue see url the catalogue functionality works pretty well on the various projects with q number composites ribo rinap rinap rinap rinap for example object type prism yield results however i have yet to perform a success catalogue search in various scores projects and sources projects these are url url url url all have local xml catalogues in their folders for example the ribo catalogue data was extracted from the ribo catalogue so the search functionality should work for both projects for example in ribo enter ruler nabopolassar or object cylinder into the search box from the page view it is clear that these should yield results but they do not either this is a bug or i do not have the correct data in the config xml files i attach the config xml for ribo and ribo sources please advise whether the config need the correct lines or whether this is a bug | 0 |
170,567 | 20,883,763,871 | IssuesEvent | 2022-03-23 01:10:12 | snowdensb/dependabot-core | https://api.github.com/repos/snowdensb/dependabot-core | reopened | WS-2021-0091 (High) detected in defaults-2.0.1.tgz | security vulnerability | ## WS-2021-0091 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>defaults-2.0.1.tgz</b></p></summary>
<p>Copies owned, enumerable properties from a source object(s) to a target object when the value of that property on the source object is `undefined`.</p>
<p>Library home page: <a href="https://registry.npmjs.org/@ndhoule/defaults/-/defaults-2.0.1.tgz">https://registry.npmjs.org/@ndhoule/defaults/-/defaults-2.0.1.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_local_file/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_local_file/node_modules/@ndhoule/defaults/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_local_file/node_modules/@ndhoule/defaults/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_local_file/node_modules/@ndhoule/defaults</p>
<p>
Dependency Hierarchy:
- analytics.js-integration-facebook-pixel-2.4.1.tgz (Root Library)
- analytics.js-integration-3.3.0.tgz
- :x: **defaults-2.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/dependabot-core/commit/ba8cd9078c8ce0cb202767d627706711237abf71">ba8cd9078c8ce0cb202767d627706711237abf71</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in `@ndhoule/defaults` versions 1.1.0 through 2.0.1 allows an attacker to cause a denial of service and may lead to remote code execution.
<p>Publish Date: 2021-05-09
<p>URL: <a href=https://github.com/ndhoule/defaults/blob/cda326293e8064ea05a3cdf05e6e82ba264722f4/index.js#L121>WS-2021-0091</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"@ndhoule/defaults","packageVersion":"2.0.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_local_file/package.json"],"isTransitiveDependency":true,"dependencyTree":"@segment/analytics.js-integration-facebook-pixel:2.4.1;@segment/analytics.js-integration:3.3.0;@ndhoule/defaults:2.0.1","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2021-0091","vulnerabilityDetails":"Prototype pollution vulnerability in `@ndhoule/defaults` versions 1.1.0 through 2.0.1 allows an attacker to cause a denial of service and may lead to remote code execution.","vulnerabilityUrl":"https://github.com/ndhoule/defaults/blob/cda326293e8064ea05a3cdf05e6e82ba264722f4/index.js#L121","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | WS-2021-0091 (High) detected in defaults-2.0.1.tgz - ## WS-2021-0091 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>defaults-2.0.1.tgz</b></p></summary>
<p>Copies owned, enumerable properties from a source object(s) to a target object when the value of that property on the source object is `undefined`.</p>
<p>Library home page: <a href="https://registry.npmjs.org/@ndhoule/defaults/-/defaults-2.0.1.tgz">https://registry.npmjs.org/@ndhoule/defaults/-/defaults-2.0.1.tgz</a></p>
<p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_local_file/package.json</p>
<p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_local_file/node_modules/@ndhoule/defaults/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_local_file/node_modules/@ndhoule/defaults/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_local_file/node_modules/@ndhoule/defaults</p>
<p>
Dependency Hierarchy:
- analytics.js-integration-facebook-pixel-2.4.1.tgz (Root Library)
- analytics.js-integration-3.3.0.tgz
- :x: **defaults-2.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/dependabot-core/commit/ba8cd9078c8ce0cb202767d627706711237abf71">ba8cd9078c8ce0cb202767d627706711237abf71</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in `@ndhoule/defaults` versions 1.1.0 through 2.0.1 allows an attacker to cause a denial of service and may lead to remote code execution.
<p>Publish Date: 2021-05-09
<p>URL: <a href=https://github.com/ndhoule/defaults/blob/cda326293e8064ea05a3cdf05e6e82ba264722f4/index.js#L121>WS-2021-0091</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"@ndhoule/defaults","packageVersion":"2.0.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_local_file/package.json"],"isTransitiveDependency":true,"dependencyTree":"@segment/analytics.js-integration-facebook-pixel:2.4.1;@segment/analytics.js-integration:3.3.0;@ndhoule/defaults:2.0.1","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2021-0091","vulnerabilityDetails":"Prototype pollution vulnerability in `@ndhoule/defaults` versions 1.1.0 through 2.0.1 allows an attacker to cause a denial of service and may lead to remote code execution.","vulnerabilityUrl":"https://github.com/ndhoule/defaults/blob/cda326293e8064ea05a3cdf05e6e82ba264722f4/index.js#L121","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | ws high detected in defaults tgz ws high severity vulnerability vulnerable library defaults tgz copies owned enumerable properties from a source object s to a target object when the value of that property on the source object is undefined library home page a href path to dependency file npm and yarn spec fixtures projects git dependency local file package json path to vulnerable library npm and yarn spec fixtures projects git dependency local file node modules ndhoule defaults package json npm and yarn spec fixtures projects git dependency local file node modules ndhoule defaults package json npm and yarn spec fixtures projects yarn git dependency local file node modules ndhoule defaults dependency hierarchy analytics js integration facebook pixel tgz root library analytics js integration tgz x defaults tgz vulnerable library found in head commit a href found in base branch main vulnerability details prototype pollution vulnerability in ndhoule defaults versions through allows an attacker to cause a denial of service and may lead to remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree segment analytics js integration facebook pixel segment analytics js integration ndhoule defaults isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails prototype pollution vulnerability in ndhoule defaults versions through allows an attacker to cause a denial of service and may lead to remote code execution vulnerabilityurl | 0 |
74,249 | 15,325,441,869 | IssuesEvent | 2021-02-26 01:19:19 | idonthaveafifaaddiction/MapLoom | https://api.github.com/repos/idonthaveafifaaddiction/MapLoom | closed | CVE-2017-16113 (High) detected in parsejson-0.0.3.tgz - autoclosed | security vulnerability | ## CVE-2017-16113 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parsejson-0.0.3.tgz</b></p></summary>
<p>Method that parses a JSON string and returns a JSON object</p>
<p>Library home page: <a href="https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz">https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz</a></p>
<p>Path to dependency file: MapLoom/vendor/bootstrap-sortable/package.json</p>
<p>Path to vulnerable library: MapLoom/vendor/bootstrap-sortable/node_modules/parsejson/package.json,MapLoom/node_modules/parsejson/package.json</p>
<p>
Dependency Hierarchy:
- karma-0.13.22.tgz (Root Library)
- socket.io-1.7.4.tgz
- socket.io-client-1.7.4.tgz
- engine.io-client-1.8.5.tgz
- :x: **parsejson-0.0.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/MapLoom/commit/58d4532ed41534d625ef81d1ac77b1b561e8870b">58d4532ed41534d625ef81d1ac77b1b561e8870b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16113>CVE-2017-16113</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"parsejson","packageVersion":"0.0.3","packageFilePaths":["/vendor/bootstrap-sortable/package.json","/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:0.13.22;socket.io:1.7.4;socket.io-client:1.7.4;engine.io-client:1.8.5;parsejson:0.0.3","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16113","vulnerabilityDetails":"The parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16113","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-16113 (High) detected in parsejson-0.0.3.tgz - autoclosed - ## CVE-2017-16113 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parsejson-0.0.3.tgz</b></p></summary>
<p>Method that parses a JSON string and returns a JSON object</p>
<p>Library home page: <a href="https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz">https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz</a></p>
<p>Path to dependency file: MapLoom/vendor/bootstrap-sortable/package.json</p>
<p>Path to vulnerable library: MapLoom/vendor/bootstrap-sortable/node_modules/parsejson/package.json,MapLoom/node_modules/parsejson/package.json</p>
<p>
Dependency Hierarchy:
- karma-0.13.22.tgz (Root Library)
- socket.io-1.7.4.tgz
- socket.io-client-1.7.4.tgz
- engine.io-client-1.8.5.tgz
- :x: **parsejson-0.0.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/MapLoom/commit/58d4532ed41534d625ef81d1ac77b1b561e8870b">58d4532ed41534d625ef81d1ac77b1b561e8870b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16113>CVE-2017-16113</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"parsejson","packageVersion":"0.0.3","packageFilePaths":["/vendor/bootstrap-sortable/package.json","/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:0.13.22;socket.io:1.7.4;socket.io-client:1.7.4;engine.io-client:1.8.5;parsejson:0.0.3","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16113","vulnerabilityDetails":"The parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16113","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in parsejson tgz autoclosed cve high severity vulnerability vulnerable library parsejson tgz method that parses a json string and returns a json object library home page a href path to dependency file maploom vendor bootstrap sortable package json path to vulnerable library maploom vendor bootstrap sortable node modules parsejson package json maploom node modules parsejson package json dependency hierarchy karma tgz root library socket io tgz socket io client tgz engine io client tgz x parsejson tgz vulnerable library found in head commit a href found in base branch master vulnerability details the parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree karma socket io socket io client engine io client parsejson isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails the parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed vulnerabilityurl | 0 |
206,484 | 15,731,680,224 | IssuesEvent | 2021-03-29 17:22:13 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | [FLAKEY TEST] contractkit-test -> contractkit -> Domain claims -> can make a domain claim | FLAKEY contractkit contractkit-test devX | Discovered at commit ea7d388fabf01a708e104e9acec99f6e05323697
Attempt No. 1:
Error: Failed to check for transaction receipt:
{}
at Object._fireError (/home/circleci/app/node_modules/web3-utils/src/index.js:57:17)
at /home/circleci/app/node_modules/web3-eth-contract/node_modules/web3-core-method/src/index.js:272:27
Attempt No. 2:
Error: Failed to check for transaction receipt:
{}
at Object._fireError (/home/circleci/app/node_modules/web3-utils/src/index.js:57:17)
at /home/circleci/app/node_modules/web3-eth-contract/node_modules/web3-core-method/src/index.js:272:27
at process._tickCallback (internal/process/next_tick.js:68:7)
Attempt No. 3:
Test Passed!
| 1.0 | [FLAKEY TEST] contractkit-test -> contractkit -> Domain claims -> can make a domain claim - Discovered at commit ea7d388fabf01a708e104e9acec99f6e05323697
Attempt No. 1:
Error: Failed to check for transaction receipt:
{}
at Object._fireError (/home/circleci/app/node_modules/web3-utils/src/index.js:57:17)
at /home/circleci/app/node_modules/web3-eth-contract/node_modules/web3-core-method/src/index.js:272:27
Attempt No. 2:
Error: Failed to check for transaction receipt:
{}
at Object._fireError (/home/circleci/app/node_modules/web3-utils/src/index.js:57:17)
at /home/circleci/app/node_modules/web3-eth-contract/node_modules/web3-core-method/src/index.js:272:27
at process._tickCallback (internal/process/next_tick.js:68:7)
Attempt No. 3:
Test Passed!
| non_priority | contractkit test contractkit domain claims can make a domain claim discovered at commit attempt no error failed to check for transaction receipt at object fireerror home circleci app node modules utils src index js at home circleci app node modules eth contract node modules core method src index js attempt no error failed to check for transaction receipt at object fireerror home circleci app node modules utils src index js at home circleci app node modules eth contract node modules core method src index js at process tickcallback internal process next tick js attempt no test passed | 0 |
306,404 | 9,392,654,150 | IssuesEvent | 2019-04-07 03:18:03 | generative-music/generative.fm | https://api.github.com/repos/generative-music/generative.fm | closed | Update indicator animation | enhancement high priority low effort | Animate the update indicator above the "About" tab label so users click on it. | 1.0 | Update indicator animation - Animate the update indicator above the "About" tab label so users click on it. | priority | update indicator animation animate the update indicator above the about tab label so users click on it | 1 |
266,489 | 8,368,326,361 | IssuesEvent | 2018-10-04 14:31:34 | andrej-szontagh/portfolio-website | https://api.github.com/repos/andrej-szontagh/portfolio-website | closed | Fix Terms&Conditions button disappearing, responsive layout / media query update. | bug priority | The layout of the scrollable text was changed but the settings for media queries wasn't so it doesn't trigger when it should .. we need to update the values. | 1.0 | Fix Terms&Conditions button disappearing, responsive layout / media query update. - The layout of the scrollable text was changed but the settings for media queries wasn't so it doesn't trigger when it should .. we need to update the values. | priority | fix terms conditions button disappearing responsive layout media query update the layout of the scrollable text was changed but the settings for media queries wasn t so it doesn t trigger when it should we need to update the values | 1 |
624,657 | 19,703,654,916 | IssuesEvent | 2022-01-12 19:18:21 | bounswe/2021SpringGroup1 | https://api.github.com/repos/bounswe/2021SpringGroup1 | closed | Writing the Final Milestone report | Priority: High Status: In Progress Platform: Android Platform: Frontend Platform: Backend | Final milestone report must be finished by tomorrow and we need to update the parts:
- RAM
- Adding final release notes
- Final project evaluation
Feel free to add some parts that I've missed by now or comment your contributions. | 1.0 | Writing the Final Milestone report - Final milestone report must be finished by tomorrow and we need to update the parts:
- RAM
- Adding final release notes
- Final project evaluation
Feel free to add some parts that I've missed by now or comment your contributions. | priority | writing the final milestone report final milestone report must be finished by tomorrow and we need to update the parts ram adding final release notes final project evaluation feel free to add some parts that i ve missed by now or comment your contributions | 1 |
827,297 | 31,765,057,160 | IssuesEvent | 2023-09-12 08:16:30 | filamentphp/filament | https://api.github.com/repos/filamentphp/filament | opened | Test | bug unconfirmed low priority | ### Package
filament/filament
### Package Version
vNothing
### Laravel Version
vNothing
### Livewire Version
vNothing
### PHP Version
vNothing
### Problem description
N/A
### Expected behavior
N/A
### Steps to reproduce
N/A
### Reproduction repository
N/A
### Relevant log output
```shell
N/A
```
| 1.0 | Test - ### Package
filament/filament
### Package Version
vNothing
### Laravel Version
vNothing
### Livewire Version
vNothing
### PHP Version
vNothing
### Problem description
N/A
### Expected behavior
N/A
### Steps to reproduce
N/A
### Reproduction repository
N/A
### Relevant log output
```shell
N/A
```
| priority | test package filament filament package version vnothing laravel version vnothing livewire version vnothing php version vnothing problem description n a expected behavior n a steps to reproduce n a reproduction repository n a relevant log output shell n a | 1 |
453,149 | 13,065,446,526 | IssuesEvent | 2020-07-30 19:49:57 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | closed | Require Parameter for hostPrefix in install.conf during cluster deployment. | High Priority scale team/ecosystem | In Scale testing there is a need to create more than 500/1000 pods per node but after creation of 510 pods there is a IP limit problem on each node for more details please refer [1] point 8 related to hostPrefix. Due to the IP limits after creation of 500 pods new pod creation will be stuck in ContainerCreation state [2]
This change should be done at the start the cluster deployment i.e. it should be changed in install-config.yaml.
[1] https://docs.openshift.com/container-platform/4.1/installing/installing_bare_metal/installing-bare-metal.html#installation-bare-metal-config-yaml_installing-bare-metal
[2] pod error from describe output
`Warning FailedCreatePodSandBox 10s (x245 over 58m) kubelet, compute-2 (combined from similar events): Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_pod-test-cephfs-26636c33118d4c3095cbd37d91ffbdcd_namespace-test-04da39dd13384c368442462a0a46b146_d78b6a81-5a2e-47fd-becf-5baf3d787ecf_0(3aaed0634517bee401dba2ae417eba78ab66cd785cb5e05460cb334f5ae895b4): Multus: [namespace-test-04da39dd13384c368442462a0a46b146/pod-test-cephfs-26636c33118d4c3095cbd37d91ffbdcd]: error adding container to network "openshift-sdn": delegateAdd: error invoking confAdd - "openshift-sdn": error in getting result from AddNetwork: CNI request failed with status 400: 'failed to run IPAM for 3aaed0634517bee401dba2ae417eba78ab66cd785cb5e05460cb334f5ae895b4: failed to run CNI IPAM ADD: failed to allocate for range 0: no IP addresses available in range set: 10.130.2.1-10.130.3.254` | 1.0 | Require Parameter for hostPrefix in install.conf during cluster deployment. - In Scale testing there is a need to create more than 500/1000 pods per node but after creation of 510 pods there is a IP limit problem on each node for more details please refer [1] point 8 related to hostPrefix. Due to the IP limits after creation of 500 pods new pod creation will be stuck in ContainerCreation state [2]
This change should be done at the start the cluster deployment i.e. it should be changed in install-config.yaml.
[1] https://docs.openshift.com/container-platform/4.1/installing/installing_bare_metal/installing-bare-metal.html#installation-bare-metal-config-yaml_installing-bare-metal
[2] pod error from describe output
`Warning FailedCreatePodSandBox 10s (x245 over 58m) kubelet, compute-2 (combined from similar events): Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_pod-test-cephfs-26636c33118d4c3095cbd37d91ffbdcd_namespace-test-04da39dd13384c368442462a0a46b146_d78b6a81-5a2e-47fd-becf-5baf3d787ecf_0(3aaed0634517bee401dba2ae417eba78ab66cd785cb5e05460cb334f5ae895b4): Multus: [namespace-test-04da39dd13384c368442462a0a46b146/pod-test-cephfs-26636c33118d4c3095cbd37d91ffbdcd]: error adding container to network "openshift-sdn": delegateAdd: error invoking confAdd - "openshift-sdn": error in getting result from AddNetwork: CNI request failed with status 400: 'failed to run IPAM for 3aaed0634517bee401dba2ae417eba78ab66cd785cb5e05460cb334f5ae895b4: failed to run CNI IPAM ADD: failed to allocate for range 0: no IP addresses available in range set: 10.130.2.1-10.130.3.254` | priority | require parameter for hostprefix in install conf during cluster deployment in scale testing there is a need to create more than pods per node but after creation of pods there is a ip limit problem on each node for more details please refer point related to hostprefix due to the ip limits after creation of pods new pod creation will be stuck in containercreation state this change should be done at the start the cluster deployment i e it should be changed in install config yaml pod error from describe output warning failedcreatepodsandbox over kubelet compute combined from similar events failed to create pod sandbox rpc error code unknown desc failed to create pod network sandbox pod test cephfs namespace test becf multus error adding container to network openshift sdn delegateadd error invoking confadd openshift sdn error in getting result from addnetwork cni request failed with status failed to run ipam for failed to run cni ipam add failed to allocate for range no ip addresses available in range set | 1 |
586,341 | 17,575,590,549 | IssuesEvent | 2021-08-15 14:47:00 | umple/umple | https://api.github.com/repos/umple/umple | closed | Certain examples do not generate graphviz class diagram with traits and methods selected | bug Component-UmpleOnline Priority-High Diffic-Med Traits | EDIT: I believe this is s special case of #1753 -- in that issue examples with traits and methods selected do not show their complete diagram; in the Decisions example, the diagram is not shown at all. Please solve #1753 first. I have not edited the remainder of the text below. I believe what is going on is that the amount of diagram shown is so small that it appears blank in the Decisions case.
## Summary
Within the UmpleOnli -- ne, the users can choose a diagram that he/she want to view. In a sample scenario, when a user selects an example ("Decisions") and wants to view the "GraphViz Class" diagram, the system can not display the diagram. It returns an empty page. Moreover, the system does not display a message to mention the reason why the diagram cannot be generated.
## Steps to Reproduce
1. Click the "Tools" menu item, select the example as "Class Diagrams >> "Decisions".

2. Now, click the "Options" menu item, select the "GraphViz Class" and wait for the diagram.

The system does not return anything. The system does not display a feedback message to mention the reason for not being able to generate the diagram.
## Expected Feature
The system should display an appropriate error message when the generation of the diagram is not possible.
A similar implementation can be found for the "GraphViz State" diagram type. | 1.0 | Certain examples do not generate graphviz class diagram with traits and methods selected - EDIT: I believe this is s special case of #1753 -- in that issue examples with traits and methods selected do not show their complete diagram; in the Decisions example, the diagram is not shown at all. Please solve #1753 first. I have not edited the remainder of the text below. I believe what is going on is that the amount of diagram shown is so small that it appears blank in the Decisions case.
## Summary
Within the UmpleOnli -- ne, the users can choose a diagram that he/she want to view. In a sample scenario, when a user selects an example ("Decisions") and wants to view the "GraphViz Class" diagram, the system can not display the diagram. It returns an empty page. Moreover, the system does not display a message to mention the reason why the diagram cannot be generated.
## Steps to Reproduce
1. Click the "Tools" menu item, select the example as "Class Diagrams >> "Decisions".

2. Now, click the "Options" menu item, select the "GraphViz Class" and wait for the diagram.

The system does not return anything. The system does not display a feedback message to mention the reason for not being able to generate the diagram.
## Expected Feature
The system should display an appropriate error message when the generation of the diagram is not possible.
A similar implementation can be found for the "GraphViz State" diagram type. | priority | certain examples do not generate graphviz class diagram with traits and methods selected edit i believe this is s special case of in that issue examples with traits and methods selected do not show their complete diagram in the decisions example the diagram is not shown at all please solve first i have not edited the remainder of the text below i believe what is going on is that the amount of diagram shown is so small that it appears blank in the decisions case summary within the umpleonli ne the users can choose a diagram that he she want to view in a sample scenario when a user selects an example decisions and wants to view the graphviz class diagram the system can not display the diagram it returns an empty page moreover the system does not display a message to mention the reason why the diagram cannot be generated steps to reproduce click the tools menu item select the example as class diagrams decisions now click the options menu item select the graphviz class and wait for the diagram the system does not return anything the system does not display a feedback message to mention the reason for not being able to generate the diagram expected feature the system should display an appropriate error message when the generation of the diagram is not possible a similar implementation can be found for the graphviz state diagram type | 1 |
326,412 | 24,083,265,335 | IssuesEvent | 2022-09-19 08:39:50 | aquarist-labs/s3gw | https://api.github.com/repos/aquarist-labs/s3gw | closed | docs: Use templates for PRs, Bugs, Issues & Feature Requests | documentation | ### What needs to be done
Create and implement an issue template for:
* Feature requests
* Issues
* Bugs
* PRs
it would need to be implemented in the following repositories:
* Main repository
* Charts
* UI
### Why it needs to be done
Currently, we're not using any template for issues or bugs. In fact, we've defined many issues in the past with barely any information on them. This makes it difficult to understand what some of these tasks are about.
implementing a simple issue template for new features & bugs should allow for more clarity when defining the scope of a task.
### Acceptance Criteria
1. A user opens an issue
2. The user is able to choose an issue template
### Additional Information
* Suggested template for new features (currently being used in this issue) https://blog.codecentric.de/en/2022/01/jira-template-user-story-tasks-bugs/
* Basic bug template: https://github.com/stevemao/github-issue-templates/blob/master/simple/ISSUE_TEMPLATE.md
| 1.0 | docs: Use templates for PRs, Bugs, Issues & Feature Requests - ### What needs to be done
Create and implement an issue template for:
* Feature requests
* Issues
* Bugs
* PRs
it would need to be implemented in the following repositories:
* Main repository
* Charts
* UI
### Why it needs to be done
Currently, we're not using any template for issues or bugs. In fact, we've defined many issues in the past with barely any information on them. This makes it difficult to understand what some of these tasks are about.
implementing a simple issue template for new features & bugs should allow for more clarity when defining the scope of a task.
### Acceptance Criteria
1. A user opens an issue
2. The user is able to choose an issue template
### Additional Information
* Suggested template for new features (currently being used in this issue) https://blog.codecentric.de/en/2022/01/jira-template-user-story-tasks-bugs/
* Basic bug template: https://github.com/stevemao/github-issue-templates/blob/master/simple/ISSUE_TEMPLATE.md
| non_priority | docs use templates for prs bugs issues feature requests what needs to be done create and implement an issue template for feature requests issues bugs prs it would need to be implemented in the following repositories main repository charts ui why it needs to be done currently we re not using any template for issues or bugs in fact we ve defined many issues in the past with barely any information on them this makes it difficult to understand what some of these tasks are about implementing a simple issue template for new features bugs should allow for more clarity when defining the scope of a task acceptance criteria a user opens an issue the user is able to choose an issue template additional information suggested template for new features currently being used in this issue basic bug template | 0 |
222,516 | 7,433,102,235 | IssuesEvent | 2018-03-26 05:51:18 | caicloud/cyclone | https://api.github.com/repos/caicloud/cyclone | closed | Generate unique name from alias when name is not provided | kind/feature priority/P0 | **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind feature
**What happened**:
**What you expected to happen**:
Can use this tool: https://github.com/mozillazg/go-slugify
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
| 1.0 | Generate unique name from alias when name is not provided - **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind feature
**What happened**:
**What you expected to happen**:
Can use this tool: https://github.com/mozillazg/go-slugify
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
| priority | generate unique name from alias when name is not provided is this a bug report or feature request kind feature what happened what you expected to happen can use this tool how to reproduce it as minimally and precisely as possible anything else we need to know | 1 |
9,027 | 23,986,166,488 | IssuesEvent | 2022-09-13 19:16:22 | GovAlta/ui-components | https://api.github.com/repos/GovAlta/ui-components | closed | Identify old "files" and communicate the DS teams intention with those files | Development Architecture Stale | This story only covers to the identification of those old files and the communication to DDI teams, not the actual decommissioning of the libraries.
**As a:** Design system team
**I want:** Cleanout "old" files related to previously supported iterations of the ui-components
**so that:** Team are less confused when they are consuming the libraries and so the DS team can get a better sense of which potential security vulnerabilities are related to the libraries in active development.
Note: dependent on the completion of #641. If we find that Vue has good support for web components we can deprecate our Vue libraries along with legacy Angular and React code.
### Acceptance criteria
- [ ] Identify "old" Angular code and any users of the code
- [ ] Identify "old" React code and any users of the code
- [ ] Identify "old" Vue code and any users of the code
- [ ] Determine how the best way to decommission this old code without affecting teams that may be consuming it currently.
- [ ] Before proceeding with the decommissioning of any code announce intentions and allow teams time to raise concerns (approx. 2 weeks before taking action)
### Presentable outcomes
- [ ] Announce the decommission of the old code 1 sprint prior to the decommissioning event.
### Sprint Ready Checklist
- Acceptance criteria defined
- Team has defined steps to satisfy the acceptance criteria
- Acceptance criteria is verifiable/testable
- External / 3rd Party dependencies identified
### Related issues:
### Resources:
| 1.0 | Identify old "files" and communicate the DS teams intention with those files - This story only covers to the identification of those old files and the communication to DDI teams, not the actual decommissioning of the libraries.
**As a:** Design system team
**I want:** Cleanout "old" files related to previously supported iterations of the ui-components
**so that:** Team are less confused when they are consuming the libraries and so the DS team can get a better sense of which potential security vulnerabilities are related to the libraries in active development.
Note: dependent on the completion of #641. If we find that Vue has good support for web components we can deprecate our Vue libraries along with legacy Angular and React code.
### Acceptance criteria
- [ ] Identify "old" Angular code and any users of the code
- [ ] Identify "old" React code and any users of the code
- [ ] Identify "old" Vue code and any users of the code
- [ ] Determine how the best way to decommission this old code without affecting teams that may be consuming it currently.
- [ ] Before proceeding with the decommissioning of any code announce intentions and allow teams time to raise concerns (approx. 2 weeks before taking action)
### Presentable outcomes
- [ ] Announce the decommission of the old code 1 sprint prior to the decommissioning event.
### Sprint Ready Checklist
- Acceptance criteria defined
- Team has defined steps to satisfy the acceptance criteria
- Acceptance criteria is verifiable/testable
- External / 3rd Party dependencies identified
### Related issues:
### Resources:
| non_priority | identify old files and communicate the ds teams intention with those files this story only covers to the identification of those old files and the communication to ddi teams not the actual decommissioning of the libraries as a design system team i want cleanout old files related to previously supported iterations of the ui components so that team are less confused when they are consuming the libraries and so the ds team can get a better sense of which potential security vulnerabilities are related to the libraries in active development note dependent on the completion of if we find that vue has good support for web components we can deprecate our vue libraries along with legacy angular and react code acceptance criteria identify old angular code and any users of the code identify old react code and any users of the code identify old vue code and any users of the code determine how the best way to decommission this old code without affecting teams that may be consuming it currently before proceeding with the decommissioning of any code announce intentions and allow teams time to raise concerns approx weeks before taking action presentable outcomes announce the decommission of the old code sprint prior to the decommissioning event sprint ready checklist acceptance criteria defined team has defined steps to satisfy the acceptance criteria acceptance criteria is verifiable testable external party dependencies identified related issues resources | 0 |
19,924 | 4,466,852,765 | IssuesEvent | 2016-08-25 00:53:30 | pgRouting/pgrouting | https://api.github.com/repos/pgRouting/pgrouting | closed | TSP documentation | Documentation Work in Progress | Bugs on documentation for the new signature of TSP
"NP-had optimization problem." should be "NP-hard optimization problem
| 1.0 | TSP documentation - Bugs on documentation for the new signature of TSP
"NP-had optimization problem." should be "NP-hard optimization problem
| non_priority | tsp documentation bugs on documentation for the new signature of tsp np had optimization problem should be np hard optimization problem | 0 |
134,303 | 19,142,553,623 | IssuesEvent | 2021-12-02 01:38:27 | RecordReplay/devtools | https://api.github.com/repos/RecordReplay/devtools | closed | Update team settings for organizations | needs-design | We need to add support for configuring organization-specific settings in the team settings pane (if the team is an org).
* Upload a team logo
* Disable "My Library"
* Disable public replays
* Set a Message of the Day
* Set Block list / Allowlist for recordings | 1.0 | Update team settings for organizations - We need to add support for configuring organization-specific settings in the team settings pane (if the team is an org).
* Upload a team logo
* Disable "My Library"
* Disable public replays
* Set a Message of the Day
* Set Block list / Allowlist for recordings | non_priority | update team settings for organizations we need to add support for configuring organization specific settings in the team settings pane if the team is an org upload a team logo disable my library disable public replays set a message of the day set block list allowlist for recordings | 0 |
788,694 | 27,761,898,603 | IssuesEvent | 2023-03-16 08:54:32 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YCQL] Fix ycqlsh cli to accept python 3.10, 3.11 by doing version comparison using float instead of string | kind/bug priority/medium area/ycql status/awaiting-triage | Jira Link: [DB-5721](https://yugabyte.atlassian.net/browse/DB-5721)
### Description
If you ONLY have python 3.10+ in your system, ycqlsh will fail to start with `No appropriate Python interpreter found.`
The reason is that it does string comparison when checking the version here: https://github.com/yugabyte/cqlsh/commit/e2cd33308aa84209216c035a688a1687254a89e7#diff-96b12a309d32d014dc9bd7ff5ee1678b47d1f361338813b26ca31af0b580142fR66
So it doesn't detect that "3.10" is a newer version than "3.6".
User on slack hitting issue on macos with only 3.11 available: https://yugabyte-db.slack.com/archives/CG0KQF0GG/p1677756783154519
[DB-5721]: https://yugabyte.atlassian.net/browse/DB-5721?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [YCQL] Fix ycqlsh cli to accept python 3.10, 3.11 by doing version comparison using float instead of string - Jira Link: [DB-5721](https://yugabyte.atlassian.net/browse/DB-5721)
### Description
If you ONLY have python 3.10+ in your system, ycqlsh will fail to start with `No appropriate Python interpreter found.`
The reason is that it does string comparison when checking the version here: https://github.com/yugabyte/cqlsh/commit/e2cd33308aa84209216c035a688a1687254a89e7#diff-96b12a309d32d014dc9bd7ff5ee1678b47d1f361338813b26ca31af0b580142fR66
So it doesn't detect that "3.10" is a newer version than "3.6".
User on slack hitting issue on macos with only 3.11 available: https://yugabyte-db.slack.com/archives/CG0KQF0GG/p1677756783154519
[DB-5721]: https://yugabyte.atlassian.net/browse/DB-5721?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | fix ycqlsh cli to accept python by doing version comparison using float instead of string jira link description if you only have python in your system ycqlsh will fail to start with no appropriate python interpreter found the reason is that it does string comparison when checking the version here so it doesn t detect that is a newer version than user on slack hitting issue on macos with only available | 1 |
130,454 | 12,429,380,880 | IssuesEvent | 2020-05-25 08:21:36 | modelica/ModelicaStandardLibrary | https://api.github.com/repos/modelica/ModelicaStandardLibrary | opened | Keyword highlighting in code segments | L: UsersGuide discussion documentation | A convention for documenting code segments (in `Modelica.UsersGuide.Conventions.Documentation.Format.Code`) states
> 5. In code segments use bold to emphasize Modelica keywords.
Does this really concerns only keywors as defined in "2.3.3 Modelica Keywords" of the Modelica Specification? In e.g. `Modelica.Fluid.UsersGuide.ComponentDefinition.UpstreamDiscretization`, there are also highlighted *inStream* operators. | 1.0 | Keyword highlighting in code segments - A convention for documenting code segments (in `Modelica.UsersGuide.Conventions.Documentation.Format.Code`) states
> 5. In code segments use bold to emphasize Modelica keywords.
Does this really concerns only keywors as defined in "2.3.3 Modelica Keywords" of the Modelica Specification? In e.g. `Modelica.Fluid.UsersGuide.ComponentDefinition.UpstreamDiscretization`, there are also highlighted *inStream* operators. | non_priority | keyword highlighting in code segments a convention for documenting code segments in modelica usersguide conventions documentation format code states in code segments use bold to emphasize modelica keywords does this really concerns only keywors as defined in modelica keywords of the modelica specification in e g modelica fluid usersguide componentdefinition upstreamdiscretization there are also highlighted instream operators | 0 |
225,453 | 24,840,108,667 | IssuesEvent | 2022-10-26 12:04:56 | helidon-io/helidon | https://api.github.com/repos/helidon-io/helidon | closed | PKCS#1 should not use internal com.sun APIs | enhancement security 4.x | To read PKCS#1 private keys, we use private API of com.sun, which is complicated and introduces additional requirements on module system. | True | PKCS#1 should not use internal com.sun APIs - To read PKCS#1 private keys, we use private API of com.sun, which is complicated and introduces additional requirements on module system. | non_priority | pkcs should not use internal com sun apis to read pkcs private keys we use private api of com sun which is complicated and introduces additional requirements on module system | 0 |
296,975 | 9,159,041,201 | IssuesEvent | 2019-03-01 00:43:17 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | Error restarting cluster: restarting kube-proxy: waiting for kube-proxy to be up for configmap update | priority/awaiting-more-evidence triage/needs-information | <!-- <HIDDEN-TEXT>
If this is a feature request, please include the expected impact to users if your feature was to be implemented. If this is a bug report, please include:
- [ ] The command-line used, including all options and flags.
- [ ] The full output of the command that failed
- [ ] The version of minikube used
- [ ] The operating system name and version used
- [ ] The output from the `minikube logs` command, if applicable.
NOTE: You may attach log files by dragging them into this window.
<END-HIDDEN-TEXT> - Please write your issue below this line. Thank you for your contribution! -->
| 1.0 | Error restarting cluster: restarting kube-proxy: waiting for kube-proxy to be up for configmap update - <!-- <HIDDEN-TEXT>
If this is a feature request, please include the expected impact to users if your feature was to be implemented. If this is a bug report, please include:
- [ ] The command-line used, including all options and flags.
- [ ] The full output of the command that failed
- [ ] The version of minikube used
- [ ] The operating system name and version used
- [ ] The output from the `minikube logs` command, if applicable.
NOTE: You may attach log files by dragging them into this window.
<END-HIDDEN-TEXT> - Please write your issue below this line. Thank you for your contribution! -->
| priority | error restarting cluster restarting kube proxy waiting for kube proxy to be up for configmap update if this is a feature request please include the expected impact to users if your feature was to be implemented if this is a bug report please include the command line used including all options and flags the full output of the command that failed the version of minikube used the operating system name and version used the output from the minikube logs command if applicable note you may attach log files by dragging them into this window please write your issue below this line thank you for your contribution | 1 |
80,104 | 3,550,776,174 | IssuesEvent | 2016-01-20 23:25:00 | apex/apex | https://api.github.com/repos/apex/apex | closed | Pass joined args as wiki topic | contrib (easy) priority refactor | Currently you'd have to do `apex wiki "Some Topic"`, but `apex wiki Some Topic` should be fine | 1.0 | Pass joined args as wiki topic - Currently you'd have to do `apex wiki "Some Topic"`, but `apex wiki Some Topic` should be fine | priority | pass joined args as wiki topic currently you d have to do apex wiki some topic but apex wiki some topic should be fine | 1 |
61,946 | 14,643,852,376 | IssuesEvent | 2020-12-25 19:10:23 | AlexRogalskiy/spark4you | https://api.github.com/repos/AlexRogalskiy/spark4you | opened | CVE-2020-13956 (Medium) detected in httpclient-4.5.6.jar | security vulnerability | ## CVE-2020-13956 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.6.jar</b></p></summary>
<p>Apache HttpComponents Client</p>
<p>Path to dependency file: spark4you/modules/spark4you-java/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.6/httpclient-4.5.6.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.6/httpclient-4.5.6.jar</p>
<p>
Dependency Hierarchy:
- spark-hive_2.12-3.0.1.jar (Root Library)
- :x: **httpclient-4.5.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark4you/commit/cf3d65692017a95100e23ebd5ace43d46c143ca5">cf3d65692017a95100e23ebd5ace43d46c143ca5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-13956 (Medium) detected in httpclient-4.5.6.jar - ## CVE-2020-13956 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.6.jar</b></p></summary>
<p>Apache HttpComponents Client</p>
<p>Path to dependency file: spark4you/modules/spark4you-java/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.6/httpclient-4.5.6.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.6/httpclient-4.5.6.jar</p>
<p>
Dependency Hierarchy:
- spark-hive_2.12-3.0.1.jar (Root Library)
- :x: **httpclient-4.5.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark4you/commit/cf3d65692017a95100e23ebd5ace43d46c143ca5">cf3d65692017a95100e23ebd5ace43d46c143ca5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in httpclient jar cve medium severity vulnerability vulnerable library httpclient jar apache httpcomponents client path to dependency file modules java pom xml path to vulnerable library home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar dependency hierarchy spark hive jar root library x httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi step up your open source security game with whitesource | 0 |
488,553 | 14,079,361,218 | IssuesEvent | 2020-11-04 14:47:19 | GSG-G9/ec-code-SOZAbit | https://api.github.com/repos/GSG-G9/ec-code-SOZAbit | closed | Buyer/ display the products | buyer priority-1 | In this section, js code will be added to handle how to display the products for the user, which includes getting the array of objects (products) from local storage to be displayed to the buyer. | 1.0 | Buyer/ display the products - In this section, js code will be added to handle how to display the products for the user, which includes getting the array of objects (products) from local storage to be displayed to the buyer. | priority | buyer display the products in this section js code will be added to handle how to display the products for the user which includes getting the array of objects products from local storage to be displayed to the buyer | 1 |
140,195 | 11,305,060,554 | IssuesEvent | 2020-01-18 02:07:10 | Comcast/trickster | https://api.github.com/repos/Comcast/trickster | opened | HTTP Compliance / Benchmarking Suite | enhancement performance tests | We would like to create a standalone compliance and performance benchmarking suite which works generically with any HTTP proxy. The idea would be that a client and server application would orchestrate and monitor HTTP requests that are routed through the interim proxy in order to time and test various requests for specification compliance and performance bottlenecks in the proxy. Solutions such as this have existed in the past, however, they appear to be commercial ($$) or abandoned. | 1.0 | HTTP Compliance / Benchmarking Suite - We would like to create a standalone compliance and performance benchmarking suite which works generically with any HTTP proxy. The idea would be that a client and server application would orchestrate and monitor HTTP requests that are routed through the interim proxy in order to time and test various requests for specification compliance and performance bottlenecks in the proxy. Solutions such as this have existed in the past, however, they appear to be commercial ($$) or abandoned. | non_priority | http compliance benchmarking suite we would like to create a standalone compliance and performance benchmarking suite which works generically with any http proxy the idea would be that a client and server application would orchestrate and monitor http requests that are routed through the interim proxy in order to time and test various requests for specification compliance and performance bottlenecks in the proxy solutions such as this have existed in the past however they appear to be commercial or abandoned | 0 |
291,352 | 25,139,553,416 | IssuesEvent | 2022-11-09 21:46:56 | astropy/astroquery | https://api.github.com/repos/astropy/astroquery | closed | Fix mocking in ALMA tests | bug testing alma | #2020 introduced the need to update the mocking in tests ` test_download_data` and `test_download_and_extract`
Without knowing anything about the data model of the downloads my workarounds for the tests looks rather hacky, so I would leave this issue to be addressed by the alma experts @keflavich and @andamian.
Bottom line: 2 remote tests are failing currently, they need to be fixed! | 1.0 | Fix mocking in ALMA tests - #2020 introduced the need to update the mocking in tests ` test_download_data` and `test_download_and_extract`
Without knowing anything about the data model of the downloads my workarounds for the tests looks rather hacky, so I would leave this issue to be addressed by the alma experts @keflavich and @andamian.
Bottom line: 2 remote tests are failing currently, they need to be fixed! | non_priority | fix mocking in alma tests introduced the need to update the mocking in tests test download data and test download and extract without knowing anything about the data model of the downloads my workarounds for the tests looks rather hacky so i would leave this issue to be addressed by the alma experts keflavich and andamian bottom line remote tests are failing currently they need to be fixed | 0 |
302,504 | 22,827,175,510 | IssuesEvent | 2022-07-12 09:36:25 | twisted/twisted | https://api.github.com/repos/twisted/twisted | closed | pydoctor generated documentation on the website should use common stylesheet | priority-normal enhancement website documentation assigned | |[<img alt="glyph's avatar" src="https://avatars.githubusercontent.com/u/716529?s=50" width="50" height="50">](https://github.com/glyph)| @glyph reported|
|-|-|
|Trac ID|trac#4693|
|Type|enhancement|
|Created|2010-10-18 23:10:04Z|
|Branch|https://github.com/twisted/twisted/tree/howto-styling-4693|
When I was setting up [tm.tl](http://tm.tl/), I realized that it's pretty easy to add the Twisted Matrix Labs template frosting to a web page. The pydoctor template that we use on the website should include that.
Attachments:
* [screen1.png](https://raw.githubusercontent.com/twisted/twistedmatrix.com-trac-attachments/trunk/ticket/e80/e8097443ce629e44b745ff723b628d3740f03141/2dbb726c073c8fbebda37819d507b352f8503c0f.png) (84552 bytes) - added by thijs on 2011-11-06 15:51:03Z -
<details><summary>Searchable metadata</summary>
```
trac-id__4693 4693
type__enhancement enhancement
reporter__glyph glyph
priority__normal normal
milestone__
branch__branches_howto_styling_4693 branches/howto-styling-4693
branch_author__thijs thijs
status__assigned assigned
resolution__None None
component__website website
keywords__documentation documentation
time__1287443404000000 1287443404000000
changetime__1321115322000000 1321115322000000
version__None None
owner__thijs thijs
cc__thijs
```
</details>
| 1.0 | pydoctor generated documentation on the website should use common stylesheet - |[<img alt="glyph's avatar" src="https://avatars.githubusercontent.com/u/716529?s=50" width="50" height="50">](https://github.com/glyph)| @glyph reported|
|-|-|
|Trac ID|trac#4693|
|Type|enhancement|
|Created|2010-10-18 23:10:04Z|
|Branch|https://github.com/twisted/twisted/tree/howto-styling-4693|
When I was setting up [tm.tl](http://tm.tl/), I realized that it's pretty easy to add the Twisted Matrix Labs template frosting to a web page. The pydoctor template that we use on the website should include that.
Attachments:
* [screen1.png](https://raw.githubusercontent.com/twisted/twistedmatrix.com-trac-attachments/trunk/ticket/e80/e8097443ce629e44b745ff723b628d3740f03141/2dbb726c073c8fbebda37819d507b352f8503c0f.png) (84552 bytes) - added by thijs on 2011-11-06 15:51:03Z -
<details><summary>Searchable metadata</summary>
```
trac-id__4693 4693
type__enhancement enhancement
reporter__glyph glyph
priority__normal normal
milestone__
branch__branches_howto_styling_4693 branches/howto-styling-4693
branch_author__thijs thijs
status__assigned assigned
resolution__None None
component__website website
keywords__documentation documentation
time__1287443404000000 1287443404000000
changetime__1321115322000000 1321115322000000
version__None None
owner__thijs thijs
cc__thijs
```
</details>
| non_priority | pydoctor generated documentation on the website should use common stylesheet glyph reported trac id trac type enhancement created branch when i was setting up i realized that it s pretty easy to add the twisted matrix labs template frosting to a web page the pydoctor template that we use on the website should include that attachments bytes added by thijs on searchable metadata trac id type enhancement enhancement reporter glyph glyph priority normal normal milestone branch branches howto styling branches howto styling branch author thijs thijs status assigned assigned resolution none none component website website keywords documentation documentation time changetime version none none owner thijs thijs cc thijs | 0 |
446,126 | 12,840,019,498 | IssuesEvent | 2020-07-07 20:18:36 | jasonlipo/contract-whist | https://api.github.com/repos/jasonlipo/contract-whist | closed | Structure of log | feature high priority | Player name should not be hard coded
Leaderboard should not be hard coded
Other ways to make smaller to fix #41 | 1.0 | Structure of log - Player name should not be hard coded
Leaderboard should not be hard coded
Other ways to make smaller to fix #41 | priority | structure of log player name should not be hard coded leaderboard should not be hard coded other ways to make smaller to fix | 1 |
54,686 | 7,918,136,409 | IssuesEvent | 2018-07-04 12:18:55 | bbc/simorgh | https://api.github.com/repos/bbc/simorgh | closed | Add documentation to subdirs | Documentation | **Is your feature request related to a problem? Please describe.**
e.g. clarify components vs containers, what we expect in each component in each of these. How we test. etc
**Describe the solution you'd like**
Keep it concise
| 1.0 | Add documentation to subdirs - **Is your feature request related to a problem? Please describe.**
e.g. clarify components vs containers, what we expect in each component in each of these. How we test. etc
**Describe the solution you'd like**
Keep it concise
| non_priority | add documentation to subdirs is your feature request related to a problem please describe e g clarify components vs containers what we expect in each component in each of these how we test etc describe the solution you d like keep it concise | 0 |
271,946 | 29,736,978,963 | IssuesEvent | 2023-06-14 02:18:03 | istio/istio | https://api.github.com/repos/istio/istio | reopened | Authz Port Matching when HBONE enabled on sidecar results in RBAC denied | area/security area/perf and scalability lifecycle/stale lifecycle/automatically-closed area/ambient | ### Bug Description
Sidecars when HBONE enabled do not have Destination Port rules apply correctly.
Steps to reproduce:
1. Install 1.16.x with `ambient` profile
2. Create two namespaces (other, test) with `istio-injection=enabled` labelled
3. Deploy sleep to `other`, httpbin to `test`
4. Apply the following Authz policy:
```yaml
apiVersion: security.istio.io/v1beta1
kind: AuthorizationPolicy
metadata:
name: httpbin
namespace: test
spec:
selector:
matchLabels:
app: httpbin
action: ALLOW
rules:
- from:
- source:
principals: ["cluster.local/ns/other/sa/sleep"]
to:
- operation:
paths:
- /get
methods: ["GET"]
ports: ["80"]
```
5. Make a curl request from sleep to httpbin's /get endpoint
Expected result:
200 response
Actual result:
403: `RBAC: access denied`
May need similar work to https://github.com/istio/istio/issues/41204
### Version
```prose
> i version
client version: 1.16.1
control plane version: 1.16.1
data plane version: 1.16.1 (4 proxies)
> k version --short
Flag --short has been deprecated, and will be removed in the future. The --short output will become the default.
Client Version: v1.26.0
Kustomize Version: v4.5.7
Server Version: v1.25.3
```
Also affects the `experimental-ambient` branch
### Additional Information
_No response_ | True | Authz Port Matching when HBONE enabled on sidecar results in RBAC denied - ### Bug Description
Sidecars when HBONE enabled do not have Destination Port rules apply correctly.
Steps to reproduce:
1. Install 1.16.x with `ambient` profile
2. Create two namespaces (other, test) with `istio-injection=enabled` labelled
3. Deploy sleep to `other`, httpbin to `test`
4. Apply the following Authz policy:
```yaml
apiVersion: security.istio.io/v1beta1
kind: AuthorizationPolicy
metadata:
name: httpbin
namespace: test
spec:
selector:
matchLabels:
app: httpbin
action: ALLOW
rules:
- from:
- source:
principals: ["cluster.local/ns/other/sa/sleep"]
to:
- operation:
paths:
- /get
methods: ["GET"]
ports: ["80"]
```
5. Make a curl request from sleep to httpbin's /get endpoint
Expected result:
200 response
Actual result:
403: `RBAC: access denied`
May need similar work to https://github.com/istio/istio/issues/41204
### Version
```prose
> i version
client version: 1.16.1
control plane version: 1.16.1
data plane version: 1.16.1 (4 proxies)
> k version --short
Flag --short has been deprecated, and will be removed in the future. The --short output will become the default.
Client Version: v1.26.0
Kustomize Version: v4.5.7
Server Version: v1.25.3
```
Also affects the `experimental-ambient` branch
### Additional Information
_No response_ | non_priority | authz port matching when hbone enabled on sidecar results in rbac denied bug description sidecars when hbone enabled do not have destination port rules apply correctly steps to reproduce install x with ambient profile create two namespaces other test with istio injection enabled labelled deploy sleep to other httpbin to test apply the following authz policy yaml apiversion security istio io kind authorizationpolicy metadata name httpbin namespace test spec selector matchlabels app httpbin action allow rules from source principals to operation paths get methods ports make a curl request from sleep to httpbin s get endpoint expected result response actual result rbac access denied may need similar work to version prose i version client version control plane version data plane version proxies k version short flag short has been deprecated and will be removed in the future the short output will become the default client version kustomize version server version also affects the experimental ambient branch additional information no response | 0 |
351,265 | 31,991,970,306 | IssuesEvent | 2023-09-21 06:39:57 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | reopened | DISABLED test_make_fx_symbolic_exhaustive_special_bessel_y1_cpu_float32 (__main__.TestProxyTensorOpInfoCPU) | triaged module: flaky-tests skipped module: fx module: unknown | Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_make_fx_symbolic_exhaustive_special_bessel_y1_cpu_float32&suite=TestProxyTensorOpInfoCPU) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/15858837799).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_make_fx_symbolic_exhaustive_special_bessel_y1_cpu_float32`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_proxy_tensor.py`
ConnectionTimeoutError: Connect timeout for 5000ms, GET https://raw.githubusercontent.com/pytorch/pytorch/main/test/test_proxy_tensor.py -2 (connected: false, keepalive socket: false, socketHandledRequests: 1, socketHandledResponses: 0)
headers: {}
cc @ezyang @SherlockNoMad @EikanWang @jgong5 @wenzhe-nrv | 1.0 | DISABLED test_make_fx_symbolic_exhaustive_special_bessel_y1_cpu_float32 (__main__.TestProxyTensorOpInfoCPU) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_make_fx_symbolic_exhaustive_special_bessel_y1_cpu_float32&suite=TestProxyTensorOpInfoCPU) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/15858837799).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_make_fx_symbolic_exhaustive_special_bessel_y1_cpu_float32`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_proxy_tensor.py`
ConnectionTimeoutError: Connect timeout for 5000ms, GET https://raw.githubusercontent.com/pytorch/pytorch/main/test/test_proxy_tensor.py -2 (connected: false, keepalive socket: false, socketHandledRequests: 1, socketHandledResponses: 0)
headers: {}
cc @ezyang @SherlockNoMad @EikanWang @jgong5 @wenzhe-nrv | non_priority | disabled test make fx symbolic exhaustive special bessel cpu main testproxytensoropinfocpu platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test make fx symbolic exhaustive special bessel cpu there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test proxy tensor py connectiontimeouterror connect timeout for get connected false keepalive socket false sockethandledrequests sockethandledresponses headers cc ezyang sherlocknomad eikanwang wenzhe nrv | 0 |
35,609 | 2,791,519,588 | IssuesEvent | 2015-05-10 06:42:47 | minj/foxtrick | https://api.github.com/repos/minj/foxtrick | closed | Consider adding 'force English flags' in CountryList | feature Misc needs-feedback Priority-Low | Original [issue 1217](https://code.google.com/p/foxtrick/issues/detail?id=1217) created by [minj](mailto:4mr.minj@gmail.com) on 2014-07-14T08:32:13.000Z:
**From:** knullig
**PostID:** [(16412248.177)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D177%26v%3D4)
**Reply:** [(16412248.1)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D1%26v%3D4)
**To:** Everyone
**Datetime:** 2014-06-26 1at 3:30
**Message:**
I like to request 2 options:
1. country names everywhere in same language
When i check the map or flaglist of the countries i have played against.
the names of the countries are all in a different language.
for example: spain is displayed as Espaรฑa
it would be great if the same name (if possible always in english) would be used everywhere.
| 1.0 | Consider adding 'force English flags' in CountryList - Original [issue 1217](https://code.google.com/p/foxtrick/issues/detail?id=1217) created by [minj](mailto:4mr.minj@gmail.com) on 2014-07-14T08:32:13.000Z:
**From:** knullig
**PostID:** [(16412248.177)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D177%26v%3D4)
**Reply:** [(16412248.1)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D1%26v%3D4)
**To:** Everyone
**Datetime:** 2014-06-26 1at 3:30
**Message:**
I like to request 2 options:
1. country names everywhere in same language
When i check the map or flaglist of the countries i have played against.
the names of the countries are all in a different language.
for example: spain is displayed as Espaรฑa
it would be great if the same name (if possible always in english) would be used everywhere.
| priority | consider adding force english flags in countrylist original created by mailto minj gmail com on from knullig postid reply to everyone datetime message i like to request options country names everywhere in same language when i check the map or flaglist of the countries i have played against the names of the countries are all in a different language for example spain is displayed as espaรฑa it would be great if the same name if possible always in english would be used everywhere | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.