Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
147,507 | 19,522,831,656 | IssuesEvent | 2021-12-29 22:28:25 | swagger-api/swagger-codegen | https://api.github.com/repos/swagger-api/swagger-codegen | opened | WS-2016-7062 (Medium) detected in plexus-utils-3.0.20.jar | security vulnerability | ## WS-2016-7062 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>plexus-utils-3.0.20.jar</b></p></summary>
<p>A collection of various utility classes to ease working with strings, files, command lines, XML and
more.</p>
<p>Path to dependency file: /modules/swagger-codegen-maven-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/plexus/plexus-utils/3.0.20/plexus-utils-3.0.20.jar</p>
<p>
Dependency Hierarchy:
- maven-core-3.2.5.jar (Root Library)
- :x: **plexus-utils-3.0.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/swagger-api/swagger-codegen/commit/4b7a8d7d7384aa6a27d6309c35ade0916edae7ed">4b7a8d7d7384aa6a27d6309c35ade0916edae7ed</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Security vulnerability found in plexus-utils before 3.0.24. XML injection found in XmlWriterUtil.java.
<p>Publish Date: 2016-05-07
<p>URL: <a href=https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de>WS-2016-7062</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de">https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de</a></p>
<p>Release Date: 2016-05-07</p>
<p>Fix Resolution: 3.0.24</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.codehaus.plexus","packageName":"plexus-utils","packageVersion":"3.0.20","packageFilePaths":["/modules/swagger-codegen-maven-plugin/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.maven:maven-core:3.2.5;org.codehaus.plexus:plexus-utils:3.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.0.24","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2016-7062","vulnerabilityDetails":"Security vulnerability found in plexus-utils before 3.0.24. XML injection found in XmlWriterUtil.java.","vulnerabilityUrl":"https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2016-7062 (Medium) detected in plexus-utils-3.0.20.jar - ## WS-2016-7062 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>plexus-utils-3.0.20.jar</b></p></summary>
<p>A collection of various utility classes to ease working with strings, files, command lines, XML and
more.</p>
<p>Path to dependency file: /modules/swagger-codegen-maven-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/plexus/plexus-utils/3.0.20/plexus-utils-3.0.20.jar</p>
<p>
Dependency Hierarchy:
- maven-core-3.2.5.jar (Root Library)
- :x: **plexus-utils-3.0.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/swagger-api/swagger-codegen/commit/4b7a8d7d7384aa6a27d6309c35ade0916edae7ed">4b7a8d7d7384aa6a27d6309c35ade0916edae7ed</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Security vulnerability found in plexus-utils before 3.0.24. XML injection found in XmlWriterUtil.java.
<p>Publish Date: 2016-05-07
<p>URL: <a href=https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de>WS-2016-7062</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de">https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de</a></p>
<p>Release Date: 2016-05-07</p>
<p>Fix Resolution: 3.0.24</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.codehaus.plexus","packageName":"plexus-utils","packageVersion":"3.0.20","packageFilePaths":["/modules/swagger-codegen-maven-plugin/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.maven:maven-core:3.2.5;org.codehaus.plexus:plexus-utils:3.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.0.24","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2016-7062","vulnerabilityDetails":"Security vulnerability found in plexus-utils before 3.0.24. XML injection found in XmlWriterUtil.java.","vulnerabilityUrl":"https://github.com/codehaus-plexus/plexus-utils/commit/f933e5e78dc2637e485447ed821fe14904f110de","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | ws medium detected in plexus utils jar ws medium severity vulnerability vulnerable library plexus utils jar a collection of various utility classes to ease working with strings files command lines xml and more path to dependency file modules swagger codegen maven plugin pom xml path to vulnerable library home wss scanner repository org codehaus plexus plexus utils plexus utils jar dependency hierarchy maven core jar root library x plexus utils jar vulnerable library found in head commit a href found in base branch master vulnerability details security vulnerability found in plexus utils before xml injection found in xmlwriterutil java publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache maven maven core org codehaus plexus plexus utils isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails security vulnerability found in plexus utils before xml injection found in xmlwriterutil java vulnerabilityurl | 0 |
441,445 | 12,718,062,109 | IssuesEvent | 2020-06-24 06:46:25 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.lowes.com - design is broken | browser-focus-geckoview engine-gecko priority-important | <!-- @browser: Firefox Mobile 77.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/54546 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.lowes.com/l/daily-deals.html?int_cmp=Home:A2:NoDivision:Merch:FW20_DOTD_Teaser
**Browser / Version**: Firefox Mobile 77.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
Nothing. Site loaded broken......
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.lowes.com - design is broken - <!-- @browser: Firefox Mobile 77.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/54546 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.lowes.com/l/daily-deals.html?int_cmp=Home:A2:NoDivision:Merch:FW20_DOTD_Teaser
**Browser / Version**: Firefox Mobile 77.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
Nothing. Site loaded broken......
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | design is broken url browser version firefox mobile operating system android tested another browser yes chrome problem type design is broken description items are overlapped steps to reproduce nothing site loaded broken browser configuration none from with ❤️ | 0 |
289,636 | 25,001,021,010 | IssuesEvent | 2022-11-03 07:55:27 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: backup/2TB/n10cpu4 failed | C-test-failure O-robot O-roachtest branch-master release-blocker | roachtest.backup/2TB/n10cpu4 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7288517?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7288517?buildTab=artifacts#/backup/2TB/n10cpu4) on master @ [9c9d55d707ad9a768027e9b7a3775c7c7cde8de7](https://github.com/cockroachdb/cockroach/commits/9c9d55d707ad9a768027e9b7a3775c7c7cde8de7):
```
test artifacts and logs in: /artifacts/backup/2TB/n10cpu4/run_1
(test_impl.go:291).Fatal: output in run_074917.293729331_n1_cockroach_sql: ./cockroach sql --insecure -e "
BACKUP bank.bank TO 'gs://cockroachdb-backup-testing/teamcity-7288517-1667452813-27-n10cpu4?AUTH=implicit'" returned: COMMAND_PROBLEM: exit status 1
(test_impl.go:291).Fatal: monitor failure: monitor task failed: t.Fatal() was called
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=true</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #91180 roachtest: backup/2TB/n10cpu4 failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-release-22.2.0 release-blocker]
- #90967 roachtest: backup/2TB/n10cpu4 failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-release-22.2]
- #80030 roachtest: backup/2TB/n10cpu4 failed [C-test-failure O-roachtest O-robot T-disaster-recovery]
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backup/2TB/n10cpu4.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: backup/2TB/n10cpu4 failed - roachtest.backup/2TB/n10cpu4 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7288517?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7288517?buildTab=artifacts#/backup/2TB/n10cpu4) on master @ [9c9d55d707ad9a768027e9b7a3775c7c7cde8de7](https://github.com/cockroachdb/cockroach/commits/9c9d55d707ad9a768027e9b7a3775c7c7cde8de7):
```
test artifacts and logs in: /artifacts/backup/2TB/n10cpu4/run_1
(test_impl.go:291).Fatal: output in run_074917.293729331_n1_cockroach_sql: ./cockroach sql --insecure -e "
BACKUP bank.bank TO 'gs://cockroachdb-backup-testing/teamcity-7288517-1667452813-27-n10cpu4?AUTH=implicit'" returned: COMMAND_PROBLEM: exit status 1
(test_impl.go:291).Fatal: monitor failure: monitor task failed: t.Fatal() was called
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=true</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #91180 roachtest: backup/2TB/n10cpu4 failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-release-22.2.0 release-blocker]
- #90967 roachtest: backup/2TB/n10cpu4 failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-release-22.2]
- #80030 roachtest: backup/2TB/n10cpu4 failed [C-test-failure O-roachtest O-robot T-disaster-recovery]
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backup/2TB/n10cpu4.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_infrastructure | roachtest backup failed roachtest backup with on master test artifacts and logs in artifacts backup run test impl go fatal output in run cockroach sql cockroach sql insecure e backup bank bank to gs cockroachdb backup testing teamcity auth implicit returned command problem exit status test impl go fatal monitor failure monitor task failed t fatal was called parameters roachtest cloud gce roachtest cpu roachtest encrypted true roachtest fs roachtest localssd true roachtest ssd help see see same failure on other branches roachtest backup failed roachtest backup failed roachtest backup failed cc cockroachdb disaster recovery | 0 |
26,971 | 20,969,709,335 | IssuesEvent | 2022-03-28 10:10:54 | florianwulfert/LogManager | https://api.github.com/repos/florianwulfert/LogManager | closed | Infrastructure: Add swagger to project | infrastructure Prio: 2 | Es soll zukünftig eine Swagger-Dokumentation geben, sodass die Schnittstellen beschreibung über Swagger gemacht wird und nicht manuell aufgeschrieben wird.
In diesem Issue soll die Initialisierung von Swagger erfolgen, sodass Swagger im Projekt genutzt werden kann. | 1.0 | Infrastructure: Add swagger to project - Es soll zukünftig eine Swagger-Dokumentation geben, sodass die Schnittstellen beschreibung über Swagger gemacht wird und nicht manuell aufgeschrieben wird.
In diesem Issue soll die Initialisierung von Swagger erfolgen, sodass Swagger im Projekt genutzt werden kann. | infrastructure | infrastructure add swagger to project es soll zukünftig eine swagger dokumentation geben sodass die schnittstellen beschreibung über swagger gemacht wird und nicht manuell aufgeschrieben wird in diesem issue soll die initialisierung von swagger erfolgen sodass swagger im projekt genutzt werden kann | 1 |
15,898 | 11,748,273,164 | IssuesEvent | 2020-03-12 14:57:37 | gap-system/gap | https://api.github.com/repos/gap-system/gap | opened | Add `~/.gap/v4.11`, `~/.gap/v4.12`, ... to default root paths? | gapdays2020-spring kind: enhancement topic: infrastructure topic: kernel | Right now, if the directory `~/.gap` exists, we automatically add it to the root paths when starting GAP, which means that users can install packages in `~/.gap/pkg`.
However, these packages then are shared by all GAP versions, and that is often not desirable. I therefore propose that we change GAP to *also* look for directories like `~/.gap/v4.11` and add those as root paths if present; of course the exact path depends on the GAP version, so in GAP 4.12.0, we'd look at `~/.gap/v4.12` instead.
I would suggest that the default order of the root paths then is this:
1. `~/.gap/v4.XYZ`
2. `~/.gap`
3. the path GAP was installed
This way, user installed packages have precedence, and those for the specific GAP version have precdence over generic ones.
I would suggest that we backport such a change to `stable-4.11` so that it will be available in GAP 4.11.1. Justification is that such a change can't really result in regressions, and making it on stable-4.11 will enable us to exploit it fully much sooner.
Moreover, ideally PackageManage would by default install into `~/.gap/v4.XYZ` instead of `~/.gap`, at least when used in GAP 4.11.1 and later.
The required kernel changes should be rather trivial, a few lines at most.
CC @mtorpey @alex-konovalov @mohamed-barakat | 1.0 | Add `~/.gap/v4.11`, `~/.gap/v4.12`, ... to default root paths? - Right now, if the directory `~/.gap` exists, we automatically add it to the root paths when starting GAP, which means that users can install packages in `~/.gap/pkg`.
However, these packages then are shared by all GAP versions, and that is often not desirable. I therefore propose that we change GAP to *also* look for directories like `~/.gap/v4.11` and add those as root paths if present; of course the exact path depends on the GAP version, so in GAP 4.12.0, we'd look at `~/.gap/v4.12` instead.
I would suggest that the default order of the root paths then is this:
1. `~/.gap/v4.XYZ`
2. `~/.gap`
3. the path GAP was installed
This way, user installed packages have precedence, and those for the specific GAP version have precdence over generic ones.
I would suggest that we backport such a change to `stable-4.11` so that it will be available in GAP 4.11.1. Justification is that such a change can't really result in regressions, and making it on stable-4.11 will enable us to exploit it fully much sooner.
Moreover, ideally PackageManage would by default install into `~/.gap/v4.XYZ` instead of `~/.gap`, at least when used in GAP 4.11.1 and later.
The required kernel changes should be rather trivial, a few lines at most.
CC @mtorpey @alex-konovalov @mohamed-barakat | infrastructure | add gap gap to default root paths right now if the directory gap exists we automatically add it to the root paths when starting gap which means that users can install packages in gap pkg however these packages then are shared by all gap versions and that is often not desirable i therefore propose that we change gap to also look for directories like gap and add those as root paths if present of course the exact path depends on the gap version so in gap we d look at gap instead i would suggest that the default order of the root paths then is this gap xyz gap the path gap was installed this way user installed packages have precedence and those for the specific gap version have precdence over generic ones i would suggest that we backport such a change to stable so that it will be available in gap justification is that such a change can t really result in regressions and making it on stable will enable us to exploit it fully much sooner moreover ideally packagemanage would by default install into gap xyz instead of gap at least when used in gap and later the required kernel changes should be rather trivial a few lines at most cc mtorpey alex konovalov mohamed barakat | 1 |
265,660 | 28,298,026,992 | IssuesEvent | 2023-04-10 01:26:27 | nk7598/linux-4.19.72 | https://api.github.com/repos/nk7598/linux-4.19.72 | closed | CVE-2020-12465 (Medium) detected in linuxlinux-4.19.269 - autoclosed | Mend: dependency security vulnerability | ## CVE-2020-12465 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.269</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An array overflow was discovered in mt76_add_fragment in drivers/net/wireless/mediatek/mt76/dma.c in the Linux kernel before 5.5.10, aka CID-b102f0c522cf. An oversized packet with too many rx fragments can corrupt memory of adjacent pages.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-12465>CVE-2020-12465</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-12465">https://www.linuxkernelcves.com/cves/CVE-2020-12465</a></p>
<p>Release Date: 2020-06-08</p>
<p>Fix Resolution: v5.6-rc6,v4.19.111,v5.4.26,v5.5.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-12465 (Medium) detected in linuxlinux-4.19.269 - autoclosed - ## CVE-2020-12465 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.269</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An array overflow was discovered in mt76_add_fragment in drivers/net/wireless/mediatek/mt76/dma.c in the Linux kernel before 5.5.10, aka CID-b102f0c522cf. An oversized packet with too many rx fragments can corrupt memory of adjacent pages.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-12465>CVE-2020-12465</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-12465">https://www.linuxkernelcves.com/cves/CVE-2020-12465</a></p>
<p>Release Date: 2020-06-08</p>
<p>Fix Resolution: v5.6-rc6,v4.19.111,v5.4.26,v5.5.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details an array overflow was discovered in add fragment in drivers net wireless mediatek dma c in the linux kernel before aka cid an oversized packet with too many rx fragments can corrupt memory of adjacent pages publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
281,705 | 8,698,426,021 | IssuesEvent | 2018-12-04 23:23:36 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | fingerprinting not blocked at all (browserprint.info, jsfiddle.net/bkf50r8v/13) | bug feature/shields/fingerprint priority/P3 | ## Description
when testing browser fingerprinting on browserprint.info (which uses a more realistic dataset than panopticlick) brave doesn't seem to be blocking anything
## Steps to Reproduce
enable fingerprint blocking globally (Block All Fingerprinting)
go to browserprint.info
click fingerprint me
fill in capcha
## Actual result:
nothing is blocked
all fonts are there
the canvas fingerprint and audio fingerprint was captured,
along with webgl data like graphics card
brave shields report 0 fingerprinting methods are blocked
## Expected result:
fingerprinting should be blocked
canvas fingerprint should be empty
graphics card should not be visible
audio fingerprint should be blocked
## Reproduces how often:
reproduces every time for me
not tested different browsers versions yet
(tested Windows only, Not tested on Muon version of Brave)
## Brave version (brave://version info)
Brave | 0.55.22 Chromium: 70.0.3538.77 (Official Build) (64-bit)
Revision | 0f6ce0b0cd63a12cb4eccea3637b1bc9a29148d9-refs/branch-heads/3538@{#1039}
### Reproducible on current release:
Yes
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields?
N/A
- Is the issue reproducible on the latest version of Chrome?
N/A
### Additional Information
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
| 1.0 | fingerprinting not blocked at all (browserprint.info, jsfiddle.net/bkf50r8v/13) - ## Description
when testing browser fingerprinting on browserprint.info (which uses a more realistic dataset than panopticlick) brave doesn't seem to be blocking anything
## Steps to Reproduce
enable fingerprint blocking globally (Block All Fingerprinting)
go to browserprint.info
click fingerprint me
fill in capcha
## Actual result:
nothing is blocked
all fonts are there
the canvas fingerprint and audio fingerprint was captured,
along with webgl data like graphics card
brave shields report 0 fingerprinting methods are blocked
## Expected result:
fingerprinting should be blocked
canvas fingerprint should be empty
graphics card should not be visible
audio fingerprint should be blocked
## Reproduces how often:
reproduces every time for me
not tested different browsers versions yet
(tested Windows only, Not tested on Muon version of Brave)
## Brave version (brave://version info)
Brave | 0.55.22 Chromium: 70.0.3538.77 (Official Build) (64-bit)
Revision | 0f6ce0b0cd63a12cb4eccea3637b1bc9a29148d9-refs/branch-heads/3538@{#1039}
### Reproducible on current release:
Yes
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields?
N/A
- Is the issue reproducible on the latest version of Chrome?
N/A
### Additional Information
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
| non_infrastructure | fingerprinting not blocked at all browserprint info jsfiddle net description when testing browser fingerprinting on browserprint info which uses a more realistic dataset than panopticlick brave doesn t seem to be blocking anything steps to reproduce enable fingerprint blocking globally block all fingerprinting go to browserprint info click fingerprint me fill in capcha actual result nothing is blocked all fonts are there the canvas fingerprint and audio fingerprint was captured along with webgl data like graphics card brave shields report fingerprinting methods are blocked expected result fingerprinting should be blocked canvas fingerprint should be empty graphics card should not be visible audio fingerprint should be blocked reproduces how often reproduces every time for me not tested different browsers versions yet tested windows only not tested on muon version of brave brave version brave version info brave chromium official build bit revision refs branch heads reproducible on current release yes website problems only does the issue resolve itself when disabling brave shields n a is the issue reproducible on the latest version of chrome n a additional information | 0 |
9,299 | 7,893,744,766 | IssuesEvent | 2018-06-28 19:04:52 | andrew-codes/SpaceMiner | https://api.github.com/repos/andrew-codes/SpaceMiner | closed | Enforce HTTPS | Infrastructure enhancement | Enforce HTTPS for the site to protect users when logging in. HTTP should redirect to HTTPS. | 1.0 | Enforce HTTPS - Enforce HTTPS for the site to protect users when logging in. HTTP should redirect to HTTPS. | infrastructure | enforce https enforce https for the site to protect users when logging in http should redirect to https | 1 |
276,000 | 30,315,320,811 | IssuesEvent | 2023-07-10 15:12:33 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | What would be the FOG endpoint when using a Failover Group listener name connecting from the Public internet | triaged cxp product-question security/subsvc Pri2 sql-managed-instance/svc | [Enter feedback here]
HI Team
The article does not mention clearly as to what would be the endpoint when using a Failover Group listener name
connecting from a Public endpoint. It's fine when there is a single SQL MI we can use <mi_name>.public.<dns_zone>.database.windows.net:3342
But what about when using a Failover Group FOG, what would be the FOG endpoint & port that we need to use to connect
from the Public internet ?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: efd9c5c9-0337-f834-4c5f-7199b5f1bc44
* Version Independent ID: 2f8374f8-28ff-c86a-38eb-5be8ea4302af
* Content: [Configure public endpoint - Azure SQL Managed Instance](https://learn.microsoft.com/en-us/azure/azure-sql/managed-instance/public-endpoint-configure?view=azuresql-mi&tabs=azure-portal)
* Content Source: [azure-sql/managed-instance/public-endpoint-configure.md](https://github.com/MicrosoftDocs/sql-docs/blob/live/azure-sql/managed-instance/public-endpoint-configure.md)
* Service: **sql-managed-instance**
* Sub-service: **security**
* GitHub Login: @zoran-rilak-msft
* Microsoft Alias: **zoranrilak** | True | What would be the FOG endpoint when using a Failover Group listener name connecting from the Public internet - [Enter feedback here]
HI Team
The article does not mention clearly as to what would be the endpoint when using a Failover Group listener name
connecting from a Public endpoint. It's fine when there is a single SQL MI we can use <mi_name>.public.<dns_zone>.database.windows.net:3342
But what about when using a Failover Group FOG, what would be the FOG endpoint & port that we need to use to connect
from the Public internet ?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: efd9c5c9-0337-f834-4c5f-7199b5f1bc44
* Version Independent ID: 2f8374f8-28ff-c86a-38eb-5be8ea4302af
* Content: [Configure public endpoint - Azure SQL Managed Instance](https://learn.microsoft.com/en-us/azure/azure-sql/managed-instance/public-endpoint-configure?view=azuresql-mi&tabs=azure-portal)
* Content Source: [azure-sql/managed-instance/public-endpoint-configure.md](https://github.com/MicrosoftDocs/sql-docs/blob/live/azure-sql/managed-instance/public-endpoint-configure.md)
* Service: **sql-managed-instance**
* Sub-service: **security**
* GitHub Login: @zoran-rilak-msft
* Microsoft Alias: **zoranrilak** | non_infrastructure | what would be the fog endpoint when using a failover group listener name connecting from the public internet hi team the article does not mention clearly as to what would be the endpoint when using a failover group listener name connecting from a public endpoint it s fine when there is a single sql mi we can use public database windows net but what about when using a failover group fog what would be the fog endpoint port that we need to use to connect from the public internet document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service sql managed instance sub service security github login zoran rilak msft microsoft alias zoranrilak | 0 |
15,639 | 11,628,073,077 | IssuesEvent | 2020-02-27 17:37:41 | enarx/enarx | https://api.github.com/repos/enarx/enarx | closed | Test for SPDX ID headers in assembly files | infrastructure | This should be a fairly straightforward adaptation of the current SPDX ID test; it just needs to check for headers in `.s` and `.S` files in addition to whatever it's doing now. | 1.0 | Test for SPDX ID headers in assembly files - This should be a fairly straightforward adaptation of the current SPDX ID test; it just needs to check for headers in `.s` and `.S` files in addition to whatever it's doing now. | infrastructure | test for spdx id headers in assembly files this should be a fairly straightforward adaptation of the current spdx id test it just needs to check for headers in s and s files in addition to whatever it s doing now | 1 |
203,178 | 15,873,528,641 | IssuesEvent | 2021-04-09 02:39:45 | amzn/selling-partner-api-docs | https://api.github.com/repos/amzn/selling-partner-api-docs | opened | Where can I find the deprecation documents for APIs? | documentation enhancement request | Hi,
As I found in [selling-partner-api-models/models](https://github.com/amzn/selling-partner-api-models/tree/main/models) directory, each models have different versions?
How can I know when will these model be deprecated?
Is there any document for them?
Thanks. | 1.0 | Where can I find the deprecation documents for APIs? - Hi,
As I found in [selling-partner-api-models/models](https://github.com/amzn/selling-partner-api-models/tree/main/models) directory, each models have different versions?
How can I know when will these model be deprecated?
Is there any document for them?
Thanks. | non_infrastructure | where can i find the deprecation documents for apis hi as i found in directory each models have different versions how can i know when will these model be deprecated is there any document for them thanks | 0 |
9,534 | 8,029,810,169 | IssuesEvent | 2018-07-27 17:19:57 | Microsoft/visualfsharp | https://api.github.com/repos/Microsoft/visualfsharp | closed | Publish new FSharp.Core patch version to include XML doc bug fix | Area-Infrastructure Area-Library | The current FSharp.Core version includes a bug in the XML docs that was fixed here: https://github.com/Microsoft/visualfsharp/commit/ee2edd15a55b3a62866004377b054eff41bb05fa
This prevents us from onboarding onto the docs.microsoft.com/dotnet/api API reference. | 1.0 | Publish new FSharp.Core patch version to include XML doc bug fix - The current FSharp.Core version includes a bug in the XML docs that was fixed here: https://github.com/Microsoft/visualfsharp/commit/ee2edd15a55b3a62866004377b054eff41bb05fa
This prevents us from onboarding onto the docs.microsoft.com/dotnet/api API reference. | infrastructure | publish new fsharp core patch version to include xml doc bug fix the current fsharp core version includes a bug in the xml docs that was fixed here this prevents us from onboarding onto the docs microsoft com dotnet api api reference | 1 |
81,419 | 10,135,584,378 | IssuesEvent | 2019-08-02 10:31:53 | vector-im/riot-web | https://api.github.com/repos/vector-im/riot-web | closed | E2E: Verifying Users: Make the member info look like the wireframe | feature needs-design type:e2e | Part of https://github.com/vector-im/riot-meta/issues/231
We still need to answer the question: 'what do we do with all the other memberinfo UX stuff we have today'? | 1.0 | E2E: Verifying Users: Make the member info look like the wireframe - Part of https://github.com/vector-im/riot-meta/issues/231
We still need to answer the question: 'what do we do with all the other memberinfo UX stuff we have today'? | non_infrastructure | verifying users make the member info look like the wireframe part of we still need to answer the question what do we do with all the other memberinfo ux stuff we have today | 0 |
97,790 | 28,502,665,233 | IssuesEvent | 2023-04-18 18:37:40 | TestCentric/testcentric-gui | https://api.github.com/repos/TestCentric/testcentric-gui | opened | Convert unit tests to use NUnitLite | Build | A long while back, I decided to use the GUI to run it's own unit tests. That works fine so long as there are no bugs, which break the execution itself! It's better to only run the GUI, with all its components, in system/package tests.
We could use nunit-console of course, but nunitlite makes it much simpler. | 1.0 | Convert unit tests to use NUnitLite - A long while back, I decided to use the GUI to run it's own unit tests. That works fine so long as there are no bugs, which break the execution itself! It's better to only run the GUI, with all its components, in system/package tests.
We could use nunit-console of course, but nunitlite makes it much simpler. | non_infrastructure | convert unit tests to use nunitlite a long while back i decided to use the gui to run it s own unit tests that works fine so long as there are no bugs which break the execution itself it s better to only run the gui with all its components in system package tests we could use nunit console of course but nunitlite makes it much simpler | 0 |
38,849 | 6,711,422,646 | IssuesEvent | 2017-10-13 03:43:52 | Pscx/Pscx | https://api.github.com/repos/Pscx/Pscx | closed | Outdated helptext for `Import-StudioVersion`'s `VisualStudioVersion` parameter | documentation | It says "Valid values are 2008, 2010, 2012 and 2013"
https://github.com/Pscx/Pscx/blob/0444d473d3dd445ad7b5fd4fc7d46f47c5362483/Src/Pscx/Modules/Utility/Pscx.Utility.psm1#L2165-L2167
when in fact the code seems to allow for many more values
https://github.com/Pscx/Pscx/blob/0444d473d3dd445ad7b5fd4fc7d46f47c5362483/Src/Pscx/Modules/Utility/Pscx.Utility.psm1#L2186-L2189
So it seems like the helptext should be updated.
Also I wonder if there's some way for `Get-Help` to read valid values from `ValidateSet` instead of having to write all the valid values in the helptext manually. | 1.0 | Outdated helptext for `Import-StudioVersion`'s `VisualStudioVersion` parameter - It says "Valid values are 2008, 2010, 2012 and 2013"
https://github.com/Pscx/Pscx/blob/0444d473d3dd445ad7b5fd4fc7d46f47c5362483/Src/Pscx/Modules/Utility/Pscx.Utility.psm1#L2165-L2167
when in fact the code seems to allow for many more values
https://github.com/Pscx/Pscx/blob/0444d473d3dd445ad7b5fd4fc7d46f47c5362483/Src/Pscx/Modules/Utility/Pscx.Utility.psm1#L2186-L2189
So it seems like the helptext should be updated.
Also I wonder if there's some way for `Get-Help` to read valid values from `ValidateSet` instead of having to write all the valid values in the helptext manually. | non_infrastructure | outdated helptext for import studioversion s visualstudioversion parameter it says valid values are and when in fact the code seems to allow for many more values so it seems like the helptext should be updated also i wonder if there s some way for get help to read valid values from validateset instead of having to write all the valid values in the helptext manually | 0 |
16,482 | 11,991,915,987 | IssuesEvent | 2020-04-08 09:12:26 | scikit-image/scikit-image | https://api.github.com/repos/scikit-image/scikit-image | closed | Broken travis CI for mac | type: infrastructure | ## Description
`+++sudo tlmgr update --self
/Library/TeX/texbin/tlmgr: unexpected return value from verify_checksum: -5`
See #4560 #4562 | 1.0 | Broken travis CI for mac - ## Description
`+++sudo tlmgr update --self
/Library/TeX/texbin/tlmgr: unexpected return value from verify_checksum: -5`
See #4560 #4562 | infrastructure | broken travis ci for mac description sudo tlmgr update self library tex texbin tlmgr unexpected return value from verify checksum see | 1 |
161,137 | 6,109,864,898 | IssuesEvent | 2017-06-21 13:58:24 | Linaro/mr-provisioner | https://api.github.com/repos/Linaro/mr-provisioner | opened | change machine bmc_info to JSONB | area/asset management difficulty/easy enhancement priority/P2 | To future-proof things, the bmc_info column, which is currently a string, should be changed to JSONB. That way, it'll be possible to save more BMC-specific information (and not lose the type of that information). For a moonshot, for example, this could be a JSON document as follows:
```javascript
{
"cartridge_id": 5
}
```
This will also make it easier to support other BMC types. | 1.0 | change machine bmc_info to JSONB - To future-proof things, the bmc_info column, which is currently a string, should be changed to JSONB. That way, it'll be possible to save more BMC-specific information (and not lose the type of that information). For a moonshot, for example, this could be a JSON document as follows:
```javascript
{
"cartridge_id": 5
}
```
This will also make it easier to support other BMC types. | non_infrastructure | change machine bmc info to jsonb to future proof things the bmc info column which is currently a string should be changed to jsonb that way it ll be possible to save more bmc specific information and not lose the type of that information for a moonshot for example this could be a json document as follows javascript cartridge id this will also make it easier to support other bmc types | 0 |
281,130 | 21,315,378,070 | IssuesEvent | 2022-04-16 07:14:06 | Graphcalibur/pe | https://api.github.com/repos/Graphcalibur/pe | opened | Caution in delete command doesn't have a proper colored box | type.DocumentationBug severity.VeryLow | 
Above is the Caution in delete command (no colored box)

Above is a Caution elsewhere in the document (has a colored box)
<!--session: 1650088173992-b9a4c6d8-5373-43e0-a30c-9dc811fe351b-->
<!--Version: Web v3.4.2--> | 1.0 | Caution in delete command doesn't have a proper colored box - 
Above is the Caution in delete command (no colored box)

Above is a Caution elsewhere in the document (has a colored box)
<!--session: 1650088173992-b9a4c6d8-5373-43e0-a30c-9dc811fe351b-->
<!--Version: Web v3.4.2--> | non_infrastructure | caution in delete command doesn t have a proper colored box above is the caution in delete command no colored box above is a caution elsewhere in the document has a colored box | 0 |
8,552 | 7,478,008,698 | IssuesEvent | 2018-04-04 10:07:44 | aokranj/website | https://api.github.com/repos/aokranj/website | closed | Create (not Vagrant but) Docker-based development environment | component:infrastructure status:3:inProgress type:feature | Fino bi blo naredit vagrant setup za development. Noter bi importal tudi dummy bazo in slike.
Kdor bi želel karkoli popravit, bi preprosto kloniral repo, pognal `vagrant up` in že bi imel working development environment.
- [ ] mysql dump
- [ ] export zadnjih 33 post slik
- [ ] export zadnje reportaže
| 1.0 | Create (not Vagrant but) Docker-based development environment - Fino bi blo naredit vagrant setup za development. Noter bi importal tudi dummy bazo in slike.
Kdor bi želel karkoli popravit, bi preprosto kloniral repo, pognal `vagrant up` in že bi imel working development environment.
- [ ] mysql dump
- [ ] export zadnjih 33 post slik
- [ ] export zadnje reportaže
| infrastructure | create not vagrant but docker based development environment fino bi blo naredit vagrant setup za development noter bi importal tudi dummy bazo in slike kdor bi želel karkoli popravit bi preprosto kloniral repo pognal vagrant up in že bi imel working development environment mysql dump export zadnjih post slik export zadnje reportaže | 1 |
79,380 | 10,121,390,922 | IssuesEvent | 2019-07-31 15:31:54 | zalando/skipper | https://api.github.com/repos/zalando/skipper | closed | Document egress specific feature | documentation | - [x] `bearerinjector` filter needs to be documented in user docs
- [x] skipper as sidecar and egress features should be documented
- [x] link from filter reference bearerinjector, circuit breaker filters to the egress feature manual with backref | 1.0 | Document egress specific feature - - [x] `bearerinjector` filter needs to be documented in user docs
- [x] skipper as sidecar and egress features should be documented
- [x] link from filter reference bearerinjector, circuit breaker filters to the egress feature manual with backref | non_infrastructure | document egress specific feature bearerinjector filter needs to be documented in user docs skipper as sidecar and egress features should be documented link from filter reference bearerinjector circuit breaker filters to the egress feature manual with backref | 0 |
8,543 | 7,473,615,391 | IssuesEvent | 2018-04-03 15:50:16 | openscholar/openscholar | https://api.github.com/repos/openscholar/openscholar | closed | Setup Alternate Varnish reverse proxy | Infrastructure | We need to replace our current application LB with a network LB so that we can make use of an assigned public IP address. A cloud formation template with an untested version of the replacement varnish stack will be sent via email. | 1.0 | Setup Alternate Varnish reverse proxy - We need to replace our current application LB with a network LB so that we can make use of an assigned public IP address. A cloud formation template with an untested version of the replacement varnish stack will be sent via email. | infrastructure | setup alternate varnish reverse proxy we need to replace our current application lb with a network lb so that we can make use of an assigned public ip address a cloud formation template with an untested version of the replacement varnish stack will be sent via email | 1 |
1,877 | 3,410,736,429 | IssuesEvent | 2015-12-04 21:37:51 | catapult-project/catapult | https://api.github.com/repos/catapult-project/catapult | closed | Need a way to denote flaky/disabled tests | Infrastructure | In #1826, we want to disable a test, but just commenting it out makes it hard to find later on. Should be some way to mark them so that we can ensure they're easy to find and follow up on. | 1.0 | Need a way to denote flaky/disabled tests - In #1826, we want to disable a test, but just commenting it out makes it hard to find later on. Should be some way to mark them so that we can ensure they're easy to find and follow up on. | infrastructure | need a way to denote flaky disabled tests in we want to disable a test but just commenting it out makes it hard to find later on should be some way to mark them so that we can ensure they re easy to find and follow up on | 1 |
29,756 | 24,251,809,983 | IssuesEvent | 2022-09-27 14:42:31 | opendatahub-io/odh-dashboard | https://api.github.com/repos/opendatahub-io/odh-dashboard | opened | [Feature Request]: Create a Migration Script INSIDE Dashboard -- Moving away from `notebookNamespace` | kind/enhancement infrastructure feature/notebook-controller priority/high | ### Feature description
In #603 we will investigate how we will move a PVC from one namespace to another. Once we have that figured out we will want to create a script that we can invoke in the Dashboard Node code to run when needed as part of a on-going migration effort.
---
- We'll need a new method (probably not a new endpoint) on the backend that can move all the related resources we create
- PVC -- See #603
- Purely spec resources (which should be as simple as copy & paste the effort to a new namespace)
- ConfigMap
- Secret
- Notebook
- We'll need a target namespace where to move them to -- this can get interesting
- There are some limitations about creating a namespace on some hostings -- so we will need to be able to smart detect which namespaces are available to the user doing the effort to see if we can move them to a specific existing namespace
- We may want the client to provide which namespace the user wants -- if they have permissions to self provision, that may also mean creating a new namespace
### Describe alternatives you've considered
_No response_
### Anything else?
Once this is in place we'll need to look at never creating in the `notebookNamespace` again. But that can be done in a cleanup ticket. | 1.0 | [Feature Request]: Create a Migration Script INSIDE Dashboard -- Moving away from `notebookNamespace` - ### Feature description
In #603 we will investigate how we will move a PVC from one namespace to another. Once we have that figured out we will want to create a script that we can invoke in the Dashboard Node code to run when needed as part of a on-going migration effort.
---
- We'll need a new method (probably not a new endpoint) on the backend that can move all the related resources we create
- PVC -- See #603
- Purely spec resources (which should be as simple as copy & paste the effort to a new namespace)
- ConfigMap
- Secret
- Notebook
- We'll need a target namespace where to move them to -- this can get interesting
- There are some limitations about creating a namespace on some hostings -- so we will need to be able to smart detect which namespaces are available to the user doing the effort to see if we can move them to a specific existing namespace
- We may want the client to provide which namespace the user wants -- if they have permissions to self provision, that may also mean creating a new namespace
### Describe alternatives you've considered
_No response_
### Anything else?
Once this is in place we'll need to look at never creating in the `notebookNamespace` again. But that can be done in a cleanup ticket. | infrastructure | create a migration script inside dashboard moving away from notebooknamespace feature description in we will investigate how we will move a pvc from one namespace to another once we have that figured out we will want to create a script that we can invoke in the dashboard node code to run when needed as part of a on going migration effort we ll need a new method probably not a new endpoint on the backend that can move all the related resources we create pvc see purely spec resources which should be as simple as copy paste the effort to a new namespace configmap secret notebook we ll need a target namespace where to move them to this can get interesting there are some limitations about creating a namespace on some hostings so we will need to be able to smart detect which namespaces are available to the user doing the effort to see if we can move them to a specific existing namespace we may want the client to provide which namespace the user wants if they have permissions to self provision that may also mean creating a new namespace describe alternatives you ve considered no response anything else once this is in place we ll need to look at never creating in the notebooknamespace again but that can be done in a cleanup ticket | 1 |
290,224 | 21,872,757,923 | IssuesEvent | 2022-05-19 07:24:25 | DjangoEx/python-engineer-roadmap | https://api.github.com/repos/DjangoEx/python-engineer-roadmap | closed | Suggestion: add a aew category | documentation enhancement | I think we can add `Social Media` as a new category for python enthusiasts.
slack, telegram, discord, etc........ groups / channels / talks/......... | 1.0 | Suggestion: add a aew category - I think we can add `Social Media` as a new category for python enthusiasts.
slack, telegram, discord, etc........ groups / channels / talks/......... | non_infrastructure | suggestion add a aew category i think we can add social media as a new category for python enthusiasts slack telegram discord etc groups channels talks | 0 |
26,783 | 20,711,531,570 | IssuesEvent | 2022-03-12 01:41:44 | microsoft/react-native-windows | https://api.github.com/repos/microsoft/react-native-windows | closed | CLI tests use outdated project files | enhancement Area: CLI Area: Tests Area: Test Infrastructure | ### Problem Description
We've made several template changes (specifically the new PackageReference work), but some of our CLI e2etests contain project files that were generated a while ago.
We should update the test assets so we can confirm that the CLI code is still being tested appropriately.
### Steps To Reproduce
yarn test
### Expected Results
_No response_
### CLI version
npx react-native --version
### Environment
```markdown
npx react-native info
```
### Target Platform Version
_No response_
### Target Device(s)
_No response_
### Visual Studio Version
_No response_
### Build Configuration
_No response_
### Snack, code example, screenshot, or link to a repository
_No response_ | 1.0 | CLI tests use outdated project files - ### Problem Description
We've made several template changes (specifically the new PackageReference work), but some of our CLI e2etests contain project files that were generated a while ago.
We should update the test assets so we can confirm that the CLI code is still being tested appropriately.
### Steps To Reproduce
yarn test
### Expected Results
_No response_
### CLI version
npx react-native --version
### Environment
```markdown
npx react-native info
```
### Target Platform Version
_No response_
### Target Device(s)
_No response_
### Visual Studio Version
_No response_
### Build Configuration
_No response_
### Snack, code example, screenshot, or link to a repository
_No response_ | infrastructure | cli tests use outdated project files problem description we ve made several template changes specifically the new packagereference work but some of our cli contain project files that were generated a while ago we should update the test assets so we can confirm that the cli code is still being tested appropriately steps to reproduce yarn test expected results no response cli version npx react native version environment markdown npx react native info target platform version no response target device s no response visual studio version no response build configuration no response snack code example screenshot or link to a repository no response | 1 |
1,370 | 3,163,834,795 | IssuesEvent | 2015-09-20 17:22:41 | Tridify/IfcOpenShell-SourceForge | https://api.github.com/repos/Tridify/IfcOpenShell-SourceForge | closed | Build ICU and Open Cascade as static libs on Windows | Infrastructure Windows/MSVC | Appears that building ICU seems rather tricky and would require Cygwin as a prerequisite. Could maybe download ICU builds from here instead http://www.npcglib.org/~stathis/blog/precompiled-icu/. | 1.0 | Build ICU and Open Cascade as static libs on Windows - Appears that building ICU seems rather tricky and would require Cygwin as a prerequisite. Could maybe download ICU builds from here instead http://www.npcglib.org/~stathis/blog/precompiled-icu/. | infrastructure | build icu and open cascade as static libs on windows appears that building icu seems rather tricky and would require cygwin as a prerequisite could maybe download icu builds from here instead | 1 |
34,410 | 29,804,128,354 | IssuesEvent | 2023-06-16 10:18:09 | DefinitelyTyped/DefinitelyTyped | https://api.github.com/repos/DefinitelyTyped/DefinitelyTyped | closed | Moving DT Issues to Discussions | Infrastructure | Hi folks, after going through the most recent user feedback for DefinitelyTyped in https://github.com/microsoft/TypeScript/issues/44030 - we came to the conclusion that moving the DefinitelyTyped issues to use the new [GitHub Discussions feature](https://docs.github.com/en/discussions) on this repo.
Issues have been an OK fit for the problem of 'this DT package has a bug', but increasingly more issues get no response and we'd like to try and re-frame how folks interact on this repo. Here's how we see it working in the future:
- "There's a bug in @types/xyz" -> Discussions
- "I'd like to request a package for xyz" -> Discussions
- "I made changes to a DT package but it didn't end up on @types/xyz" -> Issues
- "How do we handle Webpack 5 types transition?" -> Issues
Roughly, if the problem is about a single @types/xyz package then it goes into GitHub Discussions where it can be classed as 'answered' by the poster, if the problem is systemic to DefinitelyTyped - then it's in GitHub Issues.
Next Monday (June 7th) is the switch day:
- We'll have a script to close any Issues which haven't had a comment in > 6 months
- We'll change the docs in the repo and issues templates to start redirecting people to discussions
- We'll start using the bot for DT to try automate some of the notifications and tightening the format for an discussion topic to make it easier to current maintainers to get pinged | 1.0 | Moving DT Issues to Discussions - Hi folks, after going through the most recent user feedback for DefinitelyTyped in https://github.com/microsoft/TypeScript/issues/44030 - we came to the conclusion that moving the DefinitelyTyped issues to use the new [GitHub Discussions feature](https://docs.github.com/en/discussions) on this repo.
Issues have been an OK fit for the problem of 'this DT package has a bug', but increasingly more issues get no response and we'd like to try and re-frame how folks interact on this repo. Here's how we see it working in the future:
- "There's a bug in @types/xyz" -> Discussions
- "I'd like to request a package for xyz" -> Discussions
- "I made changes to a DT package but it didn't end up on @types/xyz" -> Issues
- "How do we handle Webpack 5 types transition?" -> Issues
Roughly, if the problem is about a single @types/xyz package then it goes into GitHub Discussions where it can be classed as 'answered' by the poster, if the problem is systemic to DefinitelyTyped - then it's in GitHub Issues.
Next Monday (June 7th) is the switch day:
- We'll have a script to close any Issues which haven't had a comment in > 6 months
- We'll change the docs in the repo and issues templates to start redirecting people to discussions
- We'll start using the bot for DT to try automate some of the notifications and tightening the format for an discussion topic to make it easier to current maintainers to get pinged | infrastructure | moving dt issues to discussions hi folks after going through the most recent user feedback for definitelytyped in we came to the conclusion that moving the definitelytyped issues to use the new on this repo issues have been an ok fit for the problem of this dt package has a bug but increasingly more issues get no response and we d like to try and re frame how folks interact on this repo here s how we see it working in the future there s a bug in types xyz discussions i d like to request a package for xyz discussions i made changes to a dt package but it didn t end up on types xyz issues how do we handle webpack types transition issues roughly if the problem is about a single types xyz package then it goes into github discussions where it can be classed as answered by the poster if the problem is systemic to definitelytyped then it s in github issues next monday june is the switch day we ll have a script to close any issues which haven t had a comment in months we ll change the docs in the repo and issues templates to start redirecting people to discussions we ll start using the bot for dt to try automate some of the notifications and tightening the format for an discussion topic to make it easier to current maintainers to get pinged | 1 |
252,058 | 27,226,853,363 | IssuesEvent | 2023-02-21 10:18:25 | momo-tong/spring-webmvc-new | https://api.github.com/repos/momo-tong/spring-webmvc-new | opened | groovy-all-2.4.7.jar: 2 vulnerabilities (highest severity is: 9.8) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>groovy-all-2.4.7.jar</b></p></summary>
<p>Groovy: A powerful, dynamic language for the JVM</p>
<p>Library home page: <a href="http://groovy-lang.org">http://groovy-lang.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/groovy/groovy-all/2.4.7/groovy-all-2.4.7.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/momo-tong/spring-webmvc-new/commit/8cecd370012c378c78b7c34c743e365bcd3ffd4d">8cecd370012c378c78b7c34c743e365bcd3ffd4d</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (groovy-all version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2016-6814](https://www.mend.io/vulnerability-database/CVE-2016-6814) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | groovy-all-2.4.7.jar | Direct | 2.4.8 | ❌ |
| [CVE-2020-17521](https://www.mend.io/vulnerability-database/CVE-2020-17521) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | groovy-all-2.4.7.jar | Direct | 2.4.21 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2016-6814</summary>
### Vulnerable Library - <b>groovy-all-2.4.7.jar</b></p>
<p>Groovy: A powerful, dynamic language for the JVM</p>
<p>Library home page: <a href="http://groovy-lang.org">http://groovy-lang.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/groovy/groovy-all/2.4.7/groovy-all-2.4.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **groovy-all-2.4.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/momo-tong/spring-webmvc-new/commit/8cecd370012c378c78b7c34c743e365bcd3ffd4d">8cecd370012c378c78b7c34c743e365bcd3ffd4d</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
When an application with unsupported Codehaus versions of Groovy from 1.7.0 to 2.4.3, Apache Groovy 2.4.4 to 2.4.7 on classpath uses standard Java serialization mechanisms, e.g. to communicate between servers or to store local data, it was possible for an attacker to bake a special serialized object that will execute code directly when deserialized. All applications which rely on serialization and do not isolate the code which deserializes objects were subject to this vulnerability.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6814>CVE-2016-6814</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6814">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6814</a></p>
<p>Release Date: 2018-01-15</p>
<p>Fix Resolution: 2.4.8</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-17521</summary>
### Vulnerable Library - <b>groovy-all-2.4.7.jar</b></p>
<p>Groovy: A powerful, dynamic language for the JVM</p>
<p>Library home page: <a href="http://groovy-lang.org">http://groovy-lang.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/groovy/groovy-all/2.4.7/groovy-all-2.4.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **groovy-all-2.4.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/momo-tong/spring-webmvc-new/commit/8cecd370012c378c78b7c34c743e365bcd3ffd4d">8cecd370012c378c78b7c34c743e365bcd3ffd4d</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Groovy provides extension methods to aid with creating temporary directories. Prior to this fix, Groovy's implementation of those extension methods was using a now superseded Java JDK method call that is potentially not secure on some operating systems in some contexts. Users not using the extension methods mentioned in the advisory are not affected, but may wish to read the advisory for further details. Versions Affected: 2.0 to 2.4.20, 2.5.0 to 2.5.13, 3.0.0 to 3.0.6, and 4.0.0-alpha-1. Fixed in versions 2.4.21, 2.5.14, 3.0.7, 4.0.0-alpha-2.
<p>Publish Date: 2020-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-17521>CVE-2020-17521</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/GROOVY-9824">https://issues.apache.org/jira/browse/GROOVY-9824</a></p>
<p>Release Date: 2020-12-07</p>
<p>Fix Resolution: 2.4.21</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | groovy-all-2.4.7.jar: 2 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>groovy-all-2.4.7.jar</b></p></summary>
<p>Groovy: A powerful, dynamic language for the JVM</p>
<p>Library home page: <a href="http://groovy-lang.org">http://groovy-lang.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/groovy/groovy-all/2.4.7/groovy-all-2.4.7.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/momo-tong/spring-webmvc-new/commit/8cecd370012c378c78b7c34c743e365bcd3ffd4d">8cecd370012c378c78b7c34c743e365bcd3ffd4d</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (groovy-all version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2016-6814](https://www.mend.io/vulnerability-database/CVE-2016-6814) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | groovy-all-2.4.7.jar | Direct | 2.4.8 | ❌ |
| [CVE-2020-17521](https://www.mend.io/vulnerability-database/CVE-2020-17521) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | groovy-all-2.4.7.jar | Direct | 2.4.21 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2016-6814</summary>
### Vulnerable Library - <b>groovy-all-2.4.7.jar</b></p>
<p>Groovy: A powerful, dynamic language for the JVM</p>
<p>Library home page: <a href="http://groovy-lang.org">http://groovy-lang.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/groovy/groovy-all/2.4.7/groovy-all-2.4.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **groovy-all-2.4.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/momo-tong/spring-webmvc-new/commit/8cecd370012c378c78b7c34c743e365bcd3ffd4d">8cecd370012c378c78b7c34c743e365bcd3ffd4d</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
When an application with unsupported Codehaus versions of Groovy from 1.7.0 to 2.4.3, Apache Groovy 2.4.4 to 2.4.7 on classpath uses standard Java serialization mechanisms, e.g. to communicate between servers or to store local data, it was possible for an attacker to bake a special serialized object that will execute code directly when deserialized. All applications which rely on serialization and do not isolate the code which deserializes objects were subject to this vulnerability.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6814>CVE-2016-6814</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6814">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6814</a></p>
<p>Release Date: 2018-01-15</p>
<p>Fix Resolution: 2.4.8</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-17521</summary>
### Vulnerable Library - <b>groovy-all-2.4.7.jar</b></p>
<p>Groovy: A powerful, dynamic language for the JVM</p>
<p>Library home page: <a href="http://groovy-lang.org">http://groovy-lang.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/groovy/groovy-all/2.4.7/groovy-all-2.4.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **groovy-all-2.4.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/momo-tong/spring-webmvc-new/commit/8cecd370012c378c78b7c34c743e365bcd3ffd4d">8cecd370012c378c78b7c34c743e365bcd3ffd4d</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Groovy provides extension methods to aid with creating temporary directories. Prior to this fix, Groovy's implementation of those extension methods was using a now superseded Java JDK method call that is potentially not secure on some operating systems in some contexts. Users not using the extension methods mentioned in the advisory are not affected, but may wish to read the advisory for further details. Versions Affected: 2.0 to 2.4.20, 2.5.0 to 2.5.13, 3.0.0 to 3.0.6, and 4.0.0-alpha-1. Fixed in versions 2.4.21, 2.5.14, 3.0.7, 4.0.0-alpha-2.
<p>Publish Date: 2020-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-17521>CVE-2020-17521</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/GROOVY-9824">https://issues.apache.org/jira/browse/GROOVY-9824</a></p>
<p>Release Date: 2020-12-07</p>
<p>Fix Resolution: 2.4.21</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_infrastructure | groovy all jar vulnerabilities highest severity is vulnerable library groovy all jar groovy a powerful dynamic language for the jvm library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org codehaus groovy groovy all groovy all jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in groovy all version remediation available high groovy all jar direct medium groovy all jar direct details cve vulnerable library groovy all jar groovy a powerful dynamic language for the jvm library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org codehaus groovy groovy all groovy all jar dependency hierarchy x groovy all jar vulnerable library found in head commit a href found in base branch main vulnerability details when an application with unsupported codehaus versions of groovy from to apache groovy to on classpath uses standard java serialization mechanisms e g to communicate between servers or to store local data it was possible for an attacker to bake a special serialized object that will execute code directly when deserialized all applications which rely on serialization and do not isolate the code which deserializes objects were subject to this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend cve vulnerable library groovy all jar groovy a powerful dynamic language for the jvm library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org codehaus groovy groovy all groovy all jar dependency hierarchy x groovy all jar vulnerable library found in head commit a href found in base branch main vulnerability details apache groovy provides extension methods to aid with creating temporary directories prior to this fix groovy s implementation of those extension methods was using a now superseded java jdk method call that is potentially not secure on some operating systems in some contexts users not using the extension methods mentioned in the advisory are not affected but may wish to read the advisory for further details versions affected to to to and alpha fixed in versions alpha publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
388,437 | 26,765,950,969 | IssuesEvent | 2023-01-31 10:38:20 | microsoftgraph/msgraph-developer-proxy | https://api.github.com/repos/microsoftgraph/msgraph-developer-proxy | closed | [Background Info]: Plugins Architecture | documentation | We've just merged #99 which introduces support for plugins. We should document it how it all works based on the information from the [spec](https://github.com/microsoftgraph/msgraph-developer-proxy/blob/d7147da2f29d80256da2b53ca21073da5e8c1e03/specs/plugins.md). We should also update the new structure of the config file. | 1.0 | [Background Info]: Plugins Architecture - We've just merged #99 which introduces support for plugins. We should document it how it all works based on the information from the [spec](https://github.com/microsoftgraph/msgraph-developer-proxy/blob/d7147da2f29d80256da2b53ca21073da5e8c1e03/specs/plugins.md). We should also update the new structure of the config file. | non_infrastructure | plugins architecture we ve just merged which introduces support for plugins we should document it how it all works based on the information from the we should also update the new structure of the config file | 0 |
9,815 | 8,181,702,054 | IssuesEvent | 2018-08-29 00:34:01 | dotnet/cli | https://api.github.com/repos/dotnet/cli | closed | Build failures: Unable to find package Microsoft.DotNet.Test.ProjectTemplates.2.2 | Area-Infrastructure | @dotnet-mc-bot commented on [Thu Aug 23 2018](https://github.com/dotnet/core-eng/issues/4101)
There were a set of failures during this build. Here is a summary of these:
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974073
- **Agent:** DDVSOLINUXAGE009
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974068
- **Agent:** DDVSOLINUXAGE005
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974067
- **Agent:** DDVSOLINUXAGE004
- **Error log:** Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974074
- **Agent:** DDVSOWINAGE020
- **Error log:** Unspecified error (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974069
- **Agent:** DDVSOLINUXAGE006
- **Error log:** [39;49m[36m 11>[39;49m[39;49m[31m/opt/code/build/templates/templates.csproj : error NU1102: Unable to find package Microsoft.DotNet.Test.ProjectTemplates.2.2 with version (>= 1.0.2-beta4-20180822-1972211) [/opt/code/build/...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974070
- **Agent:** DDVSOLINUXAGE007
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974075
- **Agent:** DDVSOWINAGE021
- **Error log:** Unspecified error (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974071
- **Agent:** DDVSOMACAGE011
- **Error log:** Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974072
- **Agent:** DDVSOLINUXAGE008
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
@jcagme, @markwilkie
---
@MattGal commented on [Fri Aug 24 2018](https://github.com/dotnet/core-eng/issues/4101#issuecomment-415845245)
Moving this one to CLI, @livarcocc please close if you're already tracking this. All these failures are:
```
/opt/code/build/templates/templates.csproj : error NU1102: Unable to find package Microsoft.DotNet.Test.ProjectTemplates.2.2 with version (>= 1.0.2-beta4-20180822-1972211) [/opt/code/build/BundledTemplates.proj]
/opt/code/build/templates/templates.csproj : error NU1102: - Found 2 version(s) in templating [ Nearest version: 1.0.2-beta4-20180821-1966911 ] [/opt/code/build/BundledTemplates.proj]
```
| 1.0 | Build failures: Unable to find package Microsoft.DotNet.Test.ProjectTemplates.2.2 - @dotnet-mc-bot commented on [Thu Aug 23 2018](https://github.com/dotnet/core-eng/issues/4101)
There were a set of failures during this build. Here is a summary of these:
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974073
- **Agent:** DDVSOLINUXAGE009
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974068
- **Agent:** DDVSOLINUXAGE005
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974067
- **Agent:** DDVSOLINUXAGE004
- **Error log:** Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974074
- **Agent:** DDVSOWINAGE020
- **Error log:** Unspecified error (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974069
- **Agent:** DDVSOLINUXAGE006
- **Error log:** [39;49m[36m 11>[39;49m[39;49m[31m/opt/code/build/templates/templates.csproj : error NU1102: Unable to find package Microsoft.DotNet.Test.ProjectTemplates.2.2 with version (>= 1.0.2-beta4-20180822-1972211) [/opt/code/build/...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974070
- **Agent:** DDVSOLINUXAGE007
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974075
- **Agent:** DDVSOWINAGE021
- **Error log:** Unspecified error (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974071
- **Agent:** DDVSOMACAGE011
- **Error log:** Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
* https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1974072
- **Agent:** DDVSOLINUXAGE008
- **Error log:** [?1h= Error 2147500037 (Exception from HRESULT: 0x80004005 (E_FAIL)) while compiling method BuildServerConnection.CreateMonitorDisconnectTask...
@jcagme, @markwilkie
---
@MattGal commented on [Fri Aug 24 2018](https://github.com/dotnet/core-eng/issues/4101#issuecomment-415845245)
Moving this one to CLI, @livarcocc please close if you're already tracking this. All these failures are:
```
/opt/code/build/templates/templates.csproj : error NU1102: Unable to find package Microsoft.DotNet.Test.ProjectTemplates.2.2 with version (>= 1.0.2-beta4-20180822-1972211) [/opt/code/build/BundledTemplates.proj]
/opt/code/build/templates/templates.csproj : error NU1102: - Found 2 version(s) in templating [ Nearest version: 1.0.2-beta4-20180821-1966911 ] [/opt/code/build/BundledTemplates.proj]
```
| infrastructure | build failures unable to find package microsoft dotnet test projecttemplates dotnet mc bot commented on there were a set of failures during this build here is a summary of these agent error log error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log unspecified error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log opt code build templates templates csproj error unable to find package microsoft dotnet test projecttemplates with version opt code build agent error log error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log unspecified error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask agent error log error exception from hresult e fail while compiling method buildserverconnection createmonitordisconnecttask jcagme markwilkie mattgal commented on moving this one to cli livarcocc please close if you re already tracking this all these failures are opt code build templates templates csproj error unable to find package microsoft dotnet test projecttemplates with version opt code build templates templates csproj error found version s in templating | 1 |
22,122 | 14,996,010,144 | IssuesEvent | 2021-01-29 15:02:54 | google/web-stories-wp | https://api.github.com/repos/google/web-stories-wp | closed | Update local dev environment to support PHP 8.0 | P3 PHP Pod: WP & Infra Type: Enhancement Type: Infrastructure | <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
Update local dev environment to support PHP 8.0. Either this can be done once https://github.com/docker-library/wordpress/pull/551 is merged / deployed or if we use [wp-env](https://github.com/WordPress/gutenberg/tree/master/packages/env#readme) ( #2521 / #2589 ).
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
| 1.0 | Update local dev environment to support PHP 8.0 - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
Update local dev environment to support PHP 8.0. Either this can be done once https://github.com/docker-library/wordpress/pull/551 is merged / deployed or if we use [wp-env](https://github.com/WordPress/gutenberg/tree/master/packages/env#readme) ( #2521 / #2589 ).
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
| infrastructure | update local dev environment to support php feature description update local dev environment to support php either this can be done once is merged deployed or if we use alternatives considered additional context do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief | 1 |
24,146 | 16,945,108,565 | IssuesEvent | 2021-06-28 05:12:31 | Unidata/MetPy | https://api.github.com/repos/Unidata/MetPy | closed | Upload image test results on AppVeyor | Area: Infrastructure Type: Enhancement | Looks like AppVeyor supports uploading build artifacts, so we should do so. Matplotlib's config:
```yaml
artifacts:
- path: dist\*
name: packages
- path: result_images\*
name: result_images
type: zip
on_finish:
on_failure:
- python tools/visualize_tests.py --no-browser
- echo zipping images after a failure...
- 7z a result_images.zip result_images\ | grep -v "Compressing"
- appveyor PushArtifact result_images.zip
``` | 1.0 | Upload image test results on AppVeyor - Looks like AppVeyor supports uploading build artifacts, so we should do so. Matplotlib's config:
```yaml
artifacts:
- path: dist\*
name: packages
- path: result_images\*
name: result_images
type: zip
on_finish:
on_failure:
- python tools/visualize_tests.py --no-browser
- echo zipping images after a failure...
- 7z a result_images.zip result_images\ | grep -v "Compressing"
- appveyor PushArtifact result_images.zip
``` | infrastructure | upload image test results on appveyor looks like appveyor supports uploading build artifacts so we should do so matplotlib s config yaml artifacts path dist name packages path result images name result images type zip on finish on failure python tools visualize tests py no browser echo zipping images after a failure a result images zip result images grep v compressing appveyor pushartifact result images zip | 1 |
187,019 | 6,744,347,296 | IssuesEvent | 2017-10-20 15:22:10 | Caleydo/lineupjs | https://api.github.com/repos/Caleydo/lineupjs | opened | Changing column order of a stratified column loses aggregation | priority: medium type: bug | ### Steps to reproduce
1. Stratify a column
2. Aggregate some groups
3. Drag and drop the stratified columns to a different position

### Observed behavior
The column is not stratified anymore and the aggregations are lost
### Expected behavior
* The column should remain stratified
* The aggregated groups should remain aggregated | 1.0 | Changing column order of a stratified column loses aggregation - ### Steps to reproduce
1. Stratify a column
2. Aggregate some groups
3. Drag and drop the stratified columns to a different position

### Observed behavior
The column is not stratified anymore and the aggregations are lost
### Expected behavior
* The column should remain stratified
* The aggregated groups should remain aggregated | non_infrastructure | changing column order of a stratified column loses aggregation steps to reproduce stratify a column aggregate some groups drag and drop the stratified columns to a different position observed behavior the column is not stratified anymore and the aggregations are lost expected behavior the column should remain stratified the aggregated groups should remain aggregated | 0 |
10,235 | 8,444,866,360 | IssuesEvent | 2018-10-18 19:44:07 | cmu-db/terrier | https://api.github.com/repos/cmu-db/terrier | closed | tmpsfs ramdisk on Jenkins benchmark machine | infrastructure tests | We want a tmpfs ramdisk so we can add a WAL benchmark suite to the CI runs. I followed the guide on [this page](https://www.jamescoyle.net/how-to/943-create-a-ram-disk-in-linux) when I set one up on dev4. I believe this was all I did:
```
mkdir /mnt/ramdisk
mount -t tmpfs -o size=30g tmpfs /mnt/ramdisk
```
I'll give exact ramdisk size requirements in the next 24 hours after I experiment on dev4 tonight to balance between performance variance and runtime and log file size. | 1.0 | tmpsfs ramdisk on Jenkins benchmark machine - We want a tmpfs ramdisk so we can add a WAL benchmark suite to the CI runs. I followed the guide on [this page](https://www.jamescoyle.net/how-to/943-create-a-ram-disk-in-linux) when I set one up on dev4. I believe this was all I did:
```
mkdir /mnt/ramdisk
mount -t tmpfs -o size=30g tmpfs /mnt/ramdisk
```
I'll give exact ramdisk size requirements in the next 24 hours after I experiment on dev4 tonight to balance between performance variance and runtime and log file size. | infrastructure | tmpsfs ramdisk on jenkins benchmark machine we want a tmpfs ramdisk so we can add a wal benchmark suite to the ci runs i followed the guide on when i set one up on i believe this was all i did mkdir mnt ramdisk mount t tmpfs o size tmpfs mnt ramdisk i ll give exact ramdisk size requirements in the next hours after i experiment on tonight to balance between performance variance and runtime and log file size | 1 |
179,021 | 6,620,935,277 | IssuesEvent | 2017-09-21 17:14:29 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | [Request] Add sharrre | High Priority Library - Request to Add/Update | **Library name:** sharrre
**Git repository url:** https://github.com/Julienh/Sharrre
**npm package name or url** (if there is one):
**License (List them all if it's multiple):**
**Official homepage:** http://sharrre.com/
**Wanna say something? Leave message here:**
=====================
Notes from cdnjs maintainer:
Please read the [README.md](https://github.com/cdnjs/cdnjs#cdnjs-library-repository) and [CONTRIBUTING.md](https://github.com/cdnjs/cdnjs/blob/master/CONTRIBUTING.md) document first.
We encourage you to add a library via sending pull request,
it'll be faster than just opening a request issue,
since there are tons of issues, please wait with patience,
and please don't forget to read the guidelines for contributing, thanks!!
| 1.0 | [Request] Add sharrre - **Library name:** sharrre
**Git repository url:** https://github.com/Julienh/Sharrre
**npm package name or url** (if there is one):
**License (List them all if it's multiple):**
**Official homepage:** http://sharrre.com/
**Wanna say something? Leave message here:**
=====================
Notes from cdnjs maintainer:
Please read the [README.md](https://github.com/cdnjs/cdnjs#cdnjs-library-repository) and [CONTRIBUTING.md](https://github.com/cdnjs/cdnjs/blob/master/CONTRIBUTING.md) document first.
We encourage you to add a library via sending pull request,
it'll be faster than just opening a request issue,
since there are tons of issues, please wait with patience,
and please don't forget to read the guidelines for contributing, thanks!!
| non_infrastructure | add sharrre library name sharrre git repository url npm package name or url if there is one license list them all if it s multiple official homepage wanna say something leave message here notes from cdnjs maintainer please read the and document first we encourage you to add a library via sending pull request it ll be faster than just opening a request issue since there are tons of issues please wait with patience and please don t forget to read the guidelines for contributing thanks | 0 |
30,905 | 25,153,836,695 | IssuesEvent | 2022-11-10 12:04:57 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Formatting: create tool to automate only running on files modified in a branch | area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate | Tool should do some sort of git diff --name-only to get the filenames of modified files and pipe them to jit-format.
| 1.0 | Formatting: create tool to automate only running on files modified in a branch - Tool should do some sort of git diff --name-only to get the filenames of modified files and pipe them to jit-format.
| infrastructure | formatting create tool to automate only running on files modified in a branch tool should do some sort of git diff name only to get the filenames of modified files and pipe them to jit format | 1 |
238,533 | 7,780,465,375 | IssuesEvent | 2018-06-05 20:08:39 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Cannot log back in with an account after another account was deleted - new pages | component: user profile priority: p3 size: S type: bug | STR:
1.Load AMO dev and log in with an account and go to Edit profile page
2. Delete the account - Click on Delete my profile
3. Try to log back in using another account
Expected result:
The user is logged in.
Actual result:
401 Unauthorized page is displayed.
Notes:
This is reproducible with FF60(Win10).

| 1.0 | Cannot log back in with an account after another account was deleted - new pages - STR:
1.Load AMO dev and log in with an account and go to Edit profile page
2. Delete the account - Click on Delete my profile
3. Try to log back in using another account
Expected result:
The user is logged in.
Actual result:
401 Unauthorized page is displayed.
Notes:
This is reproducible with FF60(Win10).

| non_infrastructure | cannot log back in with an account after another account was deleted new pages str load amo dev and log in with an account and go to edit profile page delete the account click on delete my profile try to log back in using another account expected result the user is logged in actual result unauthorized page is displayed notes this is reproducible with | 0 |
4,426 | 5,068,284,794 | IssuesEvent | 2016-12-24 14:40:02 | dotnet/corert | https://api.github.com/repos/dotnet/corert | closed | Cleanup the csproj files | CoreLib Infrastructure | There are a bunch of duplicated information in the csproj files.
Some of that might not be needed anymore, or not needed for all projects:
- Explicitly setting the PlatformTarget
This issue is tracking that work.
| 1.0 | Cleanup the csproj files - There are a bunch of duplicated information in the csproj files.
Some of that might not be needed anymore, or not needed for all projects:
- Explicitly setting the PlatformTarget
This issue is tracking that work.
| infrastructure | cleanup the csproj files there are a bunch of duplicated information in the csproj files some of that might not be needed anymore or not needed for all projects explicitly setting the platformtarget this issue is tracking that work | 1 |
11,732 | 9,397,812,165 | IssuesEvent | 2019-04-08 10:38:51 | elastic/beats | https://api.github.com/repos/elastic/beats | opened | [Filebeat] RabbitMQ Filebeat module | :infrastructure Filebeat in progress module | # Metricbeat Module / Dataset release checklist
This checklist is intended for Devs which create or update a module to make sure modules are consistent.
## Modules
For a fileset to go GA, the following criterias should be met:
* [ ] Supported versions are documented
* [ ] Supported operating systems are documented (if applicable)
* [ ] System tests exist
* [ ] Automated checks that all fields are documented
* [ ] Documentation
* [ ] Fields follow [ECS](https://github.com/elastic/ecs) and [naming conventions](https://www.elastic.co/guide/en/beats/devguide/master/event-conventions.html)
* [ ] Dashboards exists (if applicable)
* [ ] Kibana Home Tutorial (if applicable)
* [ ] Open issue in [EUI repo](https://github.com/elastic/eui) to add [icon for module](https://elastic.github.io/eui/#/display/icons) if not already exists.
* [ ] Open PR against Kibana repo with tutorial. Examples can be found [here](https://github.com/elastic/kibana/tree/master/src/legacy/core_plugins/kibana/server/tutorials).
## Filebeat module
* [ ] Test log files exist for the grok patterns
* [ ] Generated output for at least 1 log file exists
| 1.0 | [Filebeat] RabbitMQ Filebeat module - # Metricbeat Module / Dataset release checklist
This checklist is intended for Devs which create or update a module to make sure modules are consistent.
## Modules
For a fileset to go GA, the following criterias should be met:
* [ ] Supported versions are documented
* [ ] Supported operating systems are documented (if applicable)
* [ ] System tests exist
* [ ] Automated checks that all fields are documented
* [ ] Documentation
* [ ] Fields follow [ECS](https://github.com/elastic/ecs) and [naming conventions](https://www.elastic.co/guide/en/beats/devguide/master/event-conventions.html)
* [ ] Dashboards exists (if applicable)
* [ ] Kibana Home Tutorial (if applicable)
* [ ] Open issue in [EUI repo](https://github.com/elastic/eui) to add [icon for module](https://elastic.github.io/eui/#/display/icons) if not already exists.
* [ ] Open PR against Kibana repo with tutorial. Examples can be found [here](https://github.com/elastic/kibana/tree/master/src/legacy/core_plugins/kibana/server/tutorials).
## Filebeat module
* [ ] Test log files exist for the grok patterns
* [ ] Generated output for at least 1 log file exists
| infrastructure | rabbitmq filebeat module metricbeat module dataset release checklist this checklist is intended for devs which create or update a module to make sure modules are consistent modules for a fileset to go ga the following criterias should be met supported versions are documented supported operating systems are documented if applicable system tests exist automated checks that all fields are documented documentation fields follow and dashboards exists if applicable kibana home tutorial if applicable open issue in to add if not already exists open pr against kibana repo with tutorial examples can be found filebeat module test log files exist for the grok patterns generated output for at least log file exists | 1 |
281,562 | 21,315,415,210 | IssuesEvent | 2022-04-16 07:22:42 | atmh/pe | https://api.github.com/repos/atmh/pe | opened | UserGuide tiny images in "Listing client list by priority level of their tags" | type.DocumentationBug severity.VeryLow | The images in the section "Listing client list by priority level of their tags" of the user guide is tiny and unreadable:

<!--session: 1650087931982-ba5794c8-462a-46a9-b23b-53847f667e01-->
<!--Version: Web v3.4.2--> | 1.0 | UserGuide tiny images in "Listing client list by priority level of their tags" - The images in the section "Listing client list by priority level of their tags" of the user guide is tiny and unreadable:

<!--session: 1650087931982-ba5794c8-462a-46a9-b23b-53847f667e01-->
<!--Version: Web v3.4.2--> | non_infrastructure | userguide tiny images in listing client list by priority level of their tags the images in the section listing client list by priority level of their tags of the user guide is tiny and unreadable | 0 |
323,991 | 23,977,970,820 | IssuesEvent | 2022-09-13 13:06:25 | qir-alliance/qat | https://api.github.com/repos/qir-alliance/qat | closed | Add build support for Windows | documentation | CMake needs updating and corresponding documentation should be written. | 1.0 | Add build support for Windows - CMake needs updating and corresponding documentation should be written. | non_infrastructure | add build support for windows cmake needs updating and corresponding documentation should be written | 0 |
648,455 | 21,186,809,968 | IssuesEvent | 2022-04-08 13:31:53 | StephanAkkerman/FinTwit_Bot | https://api.github.com/repos/StephanAkkerman/FinTwit_Bot | opened | Post daily yield curve in specific channel | Priority: Low :3rd_place_medal: New feature :star: Difficulty: Medium 😐 | TradingView offers data of 1,2,3,5,7,10,20,30 year bonds. Maybe convert those into a plot like this: https://upload.wikimedia.org/wikipedia/commons/a/a7/Yield_curve_20180513.png
https://www.investopedia.com/terms/y/yieldcurve.asp For why this is important | 1.0 | Post daily yield curve in specific channel - TradingView offers data of 1,2,3,5,7,10,20,30 year bonds. Maybe convert those into a plot like this: https://upload.wikimedia.org/wikipedia/commons/a/a7/Yield_curve_20180513.png
https://www.investopedia.com/terms/y/yieldcurve.asp For why this is important | non_infrastructure | post daily yield curve in specific channel tradingview offers data of year bonds maybe convert those into a plot like this for why this is important | 0 |
9,117 | 7,834,114,632 | IssuesEvent | 2018-06-16 08:58:36 | coq/coq | https://api.github.com/repos/coq/coq | opened | [win] Current NSIS installer file is taken from patch folder - move it to /dev/nsis | kind: cleanup kind: infrastructure part: installation platform: Windows priority: low | Originally the windows build scripts where designed to download and build various versions of Coq, so it applies a few patches to Coq itself - all in the area of NSIS installer description files. These should be moved out of the patches folder to the proper place.
The folder /dev/nsis contains a completely outdated and unused version, which is way confusing. | 1.0 | [win] Current NSIS installer file is taken from patch folder - move it to /dev/nsis - Originally the windows build scripts where designed to download and build various versions of Coq, so it applies a few patches to Coq itself - all in the area of NSIS installer description files. These should be moved out of the patches folder to the proper place.
The folder /dev/nsis contains a completely outdated and unused version, which is way confusing. | infrastructure | current nsis installer file is taken from patch folder move it to dev nsis originally the windows build scripts where designed to download and build various versions of coq so it applies a few patches to coq itself all in the area of nsis installer description files these should be moved out of the patches folder to the proper place the folder dev nsis contains a completely outdated and unused version which is way confusing | 1 |
223,855 | 17,142,793,436 | IssuesEvent | 2021-07-13 11:32:44 | scylladb/scylla-monitoring | https://api.github.com/repos/scylladb/scylla-monitoring | closed | The install the monitoring without containers is out-of-date | documentation | Please make sure that this is a documentation related
install/monitor_wihtout_docker is out-of-date, it should be updated and tested with newer versions. | 1.0 | The install the monitoring without containers is out-of-date - Please make sure that this is a documentation related
install/monitor_wihtout_docker is out-of-date, it should be updated and tested with newer versions. | non_infrastructure | the install the monitoring without containers is out of date please make sure that this is a documentation related install monitor wihtout docker is out of date it should be updated and tested with newer versions | 0 |
415,122 | 12,125,159,992 | IssuesEvent | 2020-04-22 15:09:17 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Check extra parameter one service instantiation | Category: Orchestration - Flow Priority: Normal Type: Bug | **Description**
when a service with a reserved network addresses without extra parameter is instantiated, an error occurs in sinatra:
`#<NoMethodError: undefined method 'empty?' for nil:NilClass>`
**To Reproduce**
Create a service template with a reserved network like:
```
{
"name": "template-service1",
"deployment": "straight",
"description": "",
"roles": [
{
"name": "master",
"cardinality": 1,
"vm_template": 2,
"vm_template_contents": "NIC = [\n NETWORK_ID = \"$public\" ]\n",
"elasticity_policies": [],
"scheduled_policies": []
}
],
"networks": {
"public": "M|network|| |reserve_from:0:"
},
"ready_status_gate": false
}
```
Then instantiate without extra parameter.

**Expected behavior**
Clear and concise message about the error and the reason
**Details**
- Affected Component: Oneflow
- Version: development
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Check extra parameter one service instantiation - **Description**
when a service with a reserved network addresses without extra parameter is instantiated, an error occurs in sinatra:
`#<NoMethodError: undefined method 'empty?' for nil:NilClass>`
**To Reproduce**
Create a service template with a reserved network like:
```
{
"name": "template-service1",
"deployment": "straight",
"description": "",
"roles": [
{
"name": "master",
"cardinality": 1,
"vm_template": 2,
"vm_template_contents": "NIC = [\n NETWORK_ID = \"$public\" ]\n",
"elasticity_policies": [],
"scheduled_policies": []
}
],
"networks": {
"public": "M|network|| |reserve_from:0:"
},
"ready_status_gate": false
}
```
Then instantiate without extra parameter.

**Expected behavior**
Clear and concise message about the error and the reason
**Details**
- Affected Component: Oneflow
- Version: development
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| non_infrastructure | check extra parameter one service instantiation description when a service with a reserved network addresses without extra parameter is instantiated an error occurs in sinatra to reproduce create a service template with a reserved network like name template deployment straight description roles name master cardinality vm template vm template contents nic n elasticity policies scheduled policies networks public m network reserve from ready status gate false then instantiate without extra parameter expected behavior clear and concise message about the error and the reason details affected component oneflow version development additional context add any other context about the problem here progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 0 |
3,619 | 4,445,423,988 | IssuesEvent | 2016-08-20 02:19:13 | dmitrinesterenko/blog | https://api.github.com/repos/dmitrinesterenko/blog | closed | Load test homepage load | infrastructure | Load test dmitri.com to see the breaking point on the existing blog requests. | 1.0 | Load test homepage load - Load test dmitri.com to see the breaking point on the existing blog requests. | infrastructure | load test homepage load load test dmitri com to see the breaking point on the existing blog requests | 1 |
380,979 | 26,436,589,506 | IssuesEvent | 2023-01-15 13:17:45 | Seojeonguk/CSPS | https://api.github.com/repos/Seojeonguk/CSPS | closed | readme.md 업데이트 | documentation | # What
readme.md 파일 업데이트 하기
# Why
팀원 이름 변경 및 불필요한 접기 파일 제거, 기술 스택 뱃지 추가
# How
- [ ] 팀원 멘션 이름 변경(LEE HAE JINN -> JIN HAE JIN)
- [ ] 기술 스택 접기 제거
- [ ] 상단 기술 스택 뱃지 추가 | 1.0 | readme.md 업데이트 - # What
readme.md 파일 업데이트 하기
# Why
팀원 이름 변경 및 불필요한 접기 파일 제거, 기술 스택 뱃지 추가
# How
- [ ] 팀원 멘션 이름 변경(LEE HAE JINN -> JIN HAE JIN)
- [ ] 기술 스택 접기 제거
- [ ] 상단 기술 스택 뱃지 추가 | non_infrastructure | readme md 업데이트 what readme md 파일 업데이트 하기 why 팀원 이름 변경 및 불필요한 접기 파일 제거 기술 스택 뱃지 추가 how 팀원 멘션 이름 변경 lee hae jinn jin hae jin 기술 스택 접기 제거 상단 기술 스택 뱃지 추가 | 0 |
121,582 | 17,661,269,103 | IssuesEvent | 2021-08-21 14:59:04 | turkdevops/docs-2 | https://api.github.com/repos/turkdevops/docs-2 | opened | CVE-2020-28469 (High) detected in multiple libraries | security vulnerability | ## CVE-2020-28469 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-5.1.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-4.44.0.tgz (Root Library)
- watchpack-1.7.4.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: docs-2/package.json</p>
<p>Path to vulnerable library: docs-2/node_modules/copy-webpack-plugin/node_modules/glob-parent/package.json,docs-2/node_modules/chokidar/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-2.0.4.tgz (Root Library)
- chokidar-3.4.1.tgz
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.0.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.0.tgz</a></p>
<p>Path to dependency file: docs-2/package.json</p>
<p>Path to vulnerable library: docs-2/node_modules/eslint/node_modules/glob-parent/package.json,docs-2/node_modules/fast-glob/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- standard-14.3.1.tgz (Root Library)
- eslint-6.4.0.tgz
- :x: **glob-parent-5.1.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/docs-2/commit/c6d22df020239a555599b5225d22e6a61ddea846">c6d22df020239a555599b5225d22e6a61ddea846</a></p>
<p>Found in base branch: <b>remove-starter-workflow</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-28469 (High) detected in multiple libraries - ## CVE-2020-28469 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-5.1.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-4.44.0.tgz (Root Library)
- watchpack-1.7.4.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: docs-2/package.json</p>
<p>Path to vulnerable library: docs-2/node_modules/copy-webpack-plugin/node_modules/glob-parent/package.json,docs-2/node_modules/chokidar/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-2.0.4.tgz (Root Library)
- chokidar-3.4.1.tgz
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.0.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.0.tgz</a></p>
<p>Path to dependency file: docs-2/package.json</p>
<p>Path to vulnerable library: docs-2/node_modules/eslint/node_modules/glob-parent/package.json,docs-2/node_modules/fast-glob/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- standard-14.3.1.tgz (Root Library)
- eslint-6.4.0.tgz
- :x: **glob-parent-5.1.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/docs-2/commit/c6d22df020239a555599b5225d22e6a61ddea846">c6d22df020239a555599b5225d22e6a61ddea846</a></p>
<p>Found in base branch: <b>remove-starter-workflow</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries glob parent tgz glob parent tgz glob parent tgz glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href dependency hierarchy webpack tgz root library watchpack tgz watchpack tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file docs package json path to vulnerable library docs node modules copy webpack plugin node modules glob parent package json docs node modules chokidar node modules glob parent package json dependency hierarchy nodemon tgz root library chokidar tgz x glob parent tgz vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file docs package json path to vulnerable library docs node modules eslint node modules glob parent package json docs node modules fast glob node modules glob parent package json dependency hierarchy standard tgz root library eslint tgz x glob parent tgz vulnerable library found in head commit a href found in base branch remove starter workflow vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource | 0 |
48,505 | 20,161,471,866 | IssuesEvent | 2022-02-09 22:05:28 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | Az apim check-name failing | Service Attention API Management customer-reported needs-team-attention |
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az apim check-name`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
No module named 'azure.mgmt.apimanagement.models.api_management_client_enums'
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\knack\knack\cli.py", line 206, in invoke
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli-core\azure\cli\core\commands\__init__.py", line 528, in execute
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli-core\azure\cli\core\__init__.py", line 300, in load_arguments
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli\azure\cli\command_modules\apim\__init__.py", line 28, in load_arguments
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli\azure\cli\command_modules\apim\_params.py", line 11, in <module>
ModuleNotFoundError: No module named 'azure.mgmt.apimanagement.models.api_management_client_enums'
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az apim check-name --name {}`
## Expected Behavior
## Environment Summary
```
Windows-10-10.0.18362-SP0
Python 3.6.6
azure-cli 2.1.0
Extensions:
interactive 0.4.3
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
| 1.0 | Az apim check-name failing -
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az apim check-name`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
No module named 'azure.mgmt.apimanagement.models.api_management_client_enums'
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\knack\knack\cli.py", line 206, in invoke
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli-core\azure\cli\core\commands\__init__.py", line 528, in execute
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli-core\azure\cli\core\__init__.py", line 300, in load_arguments
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli\azure\cli\command_modules\apim\__init__.py", line 28, in load_arguments
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-3pnsb50r\azure-cli\azure\cli\command_modules\apim\_params.py", line 11, in <module>
ModuleNotFoundError: No module named 'azure.mgmt.apimanagement.models.api_management_client_enums'
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az apim check-name --name {}`
## Expected Behavior
## Environment Summary
```
Windows-10-10.0.18362-SP0
Python 3.6.6
azure-cli 2.1.0
Extensions:
interactive 0.4.3
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
| non_infrastructure | az apim check name failing this is autogenerated please review and update as needed describe the bug command name az apim check name errors the command failed with an unexpected error here is the traceback no module named azure mgmt apimanagement models api management client enums traceback most recent call last file c users vssadm appdata local temp pip install knack knack cli py line in invoke file c users vssadm appdata local temp pip install azure cli core azure cli core commands init py line in execute file c users vssadm appdata local temp pip install azure cli core azure cli core init py line in load arguments file c users vssadm appdata local temp pip install azure cli azure cli command modules apim init py line in load arguments file c users vssadm appdata local temp pip install azure cli azure cli command modules apim params py line in modulenotfounderror no module named azure mgmt apimanagement models api management client enums to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az apim check name name expected behavior environment summary windows python azure cli extensions interactive additional context | 0 |
77,356 | 7,571,996,560 | IssuesEvent | 2018-04-23 13:52:06 | tum-gis/3dcitydb-docker-postgis | https://api.github.com/repos/tum-gis/3dcitydb-docker-postgis | closed | Invalid SRID | test | Test behavior, if invalid SRID is provided with `docker run`. E.g. if the provided SRID is unknown to the 3DCityDB. | 1.0 | Invalid SRID - Test behavior, if invalid SRID is provided with `docker run`. E.g. if the provided SRID is unknown to the 3DCityDB. | non_infrastructure | invalid srid test behavior if invalid srid is provided with docker run e g if the provided srid is unknown to the | 0 |
10,148 | 31,810,023,253 | IssuesEvent | 2023-09-13 16:10:49 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | closed | Adjust ui tests to Object Storage ui elements | ui_automation Squad/Black | 
4.14 ODF deployment have changes so we need to adjust PageNavigator and MCG, Object related tests | 1.0 | Adjust ui tests to Object Storage ui elements - 
4.14 ODF deployment have changes so we need to adjust PageNavigator and MCG, Object related tests | non_infrastructure | adjust ui tests to object storage ui elements odf deployment have changes so we need to adjust pagenavigator and mcg object related tests | 0 |
10,496 | 8,585,426,118 | IssuesEvent | 2018-11-14 03:17:12 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | query operator with JMESPath doesn't handle a 'dash' (-) properly | Infrastructure bug | If a tag on a resource has a name that includes a `-` the query fails incorrectly.
**To Reproduce**
1. Assign a tag of `CREATED-BY` to a resource group in Azure with a value of 'foobar'
2. Run the Query
`az group list -o tsv --query "[?tags.CREATED-BY == 'foobar']"`
**Expected behavior**
Normally the results will show the item.
**Actual result**
The error shows as
```
az group list: error: argument --query: invalid jmespath_type value: "[?tags.CREATED-BY == 'foobar']"
```
If you run a similar query on http://jmespath.org/ as below, it works fine
```
locations[?state == 'WA' && foo-bar == 'foobar'].name | sort(@)
```
Using data of:
```json
{
"locations": [
{"name": "Seattssle", "state": "WA", "foo-bar": "foobar"},
{"name": "New York", "state": "NY"},
{"name": "Bellevue", "state": "WA"},
{"name": "Olympia", "state": "WA"}
]
}
```
**Environment summary**
This is verified to fail in the Azure Cloud Shell.
**Additional context**
Add any other context about the problem here.
| 1.0 | query operator with JMESPath doesn't handle a 'dash' (-) properly - If a tag on a resource has a name that includes a `-` the query fails incorrectly.
**To Reproduce**
1. Assign a tag of `CREATED-BY` to a resource group in Azure with a value of 'foobar'
2. Run the Query
`az group list -o tsv --query "[?tags.CREATED-BY == 'foobar']"`
**Expected behavior**
Normally the results will show the item.
**Actual result**
The error shows as
```
az group list: error: argument --query: invalid jmespath_type value: "[?tags.CREATED-BY == 'foobar']"
```
If you run a similar query on http://jmespath.org/ as below, it works fine
```
locations[?state == 'WA' && foo-bar == 'foobar'].name | sort(@)
```
Using data of:
```json
{
"locations": [
{"name": "Seattssle", "state": "WA", "foo-bar": "foobar"},
{"name": "New York", "state": "NY"},
{"name": "Bellevue", "state": "WA"},
{"name": "Olympia", "state": "WA"}
]
}
```
**Environment summary**
This is verified to fail in the Azure Cloud Shell.
**Additional context**
Add any other context about the problem here.
| infrastructure | query operator with jmespath doesn t handle a dash properly if a tag on a resource has a name that includes a the query fails incorrectly to reproduce assign a tag of created by to a resource group in azure with a value of foobar run the query az group list o tsv query expected behavior normally the results will show the item actual result the error shows as az group list error argument query invalid jmespath type value if you run a similar query on as below it works fine locations name sort using data of json locations name seattssle state wa foo bar foobar name new york state ny name bellevue state wa name olympia state wa environment summary this is verified to fail in the azure cloud shell additional context add any other context about the problem here | 1 |
103,390 | 16,602,496,458 | IssuesEvent | 2021-06-01 21:41:01 | gms-ws-sandbox/nibrs | https://api.github.com/repos/gms-ws-sandbox/nibrs | opened | CVE-2020-35491 (High) detected in multiple libraries | security vulnerability | ## CVE-2020-35491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-validate-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-summary-report-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.1.5.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-validation/pom.xml</p>
<p>Path to vulnerable library: nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/dba6b0930aa319c568021490e9259f5cae89b6c5">dba6b0930aa319c568021490e9259f5cae89b6c5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-validation/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-summary-report/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-35491","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-35491 (High) detected in multiple libraries - ## CVE-2020-35491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-validate-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-summary-report-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.1.5.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-validation/pom.xml</p>
<p>Path to vulnerable library: nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/dba6b0930aa319c568021490e9259f5cae89b6c5">dba6b0930aa319c568021490e9259f5cae89b6c5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-validation/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-summary-report/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-35491","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs validate common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs validation pom xml path to vulnerable library nibrs web nibrs web target nibrs web web inf lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources sharedpooldatasource vulnerabilityurl | 0 |
9,039 | 7,781,390,313 | IssuesEvent | 2018-06-05 23:57:21 | astropy/astroquery | https://api.github.com/repos/astropy/astroquery | opened | Continuous Release: remove 'dev'? | infrastructure | The continuous release process was established in #1124, but it doesn't fully "work" yet because we're pushing .dev releases to pypi, which means that `pip install astroquery` doesn't install the latest version of astroquery.
So, what is the appropriate way to make this a fully automated release process? Can we just drop the 'dev' from the version name? Does that break some of the astropy-helpers machinery?
cc @eteq, @bsipocz | 1.0 | Continuous Release: remove 'dev'? - The continuous release process was established in #1124, but it doesn't fully "work" yet because we're pushing .dev releases to pypi, which means that `pip install astroquery` doesn't install the latest version of astroquery.
So, what is the appropriate way to make this a fully automated release process? Can we just drop the 'dev' from the version name? Does that break some of the astropy-helpers machinery?
cc @eteq, @bsipocz | infrastructure | continuous release remove dev the continuous release process was established in but it doesn t fully work yet because we re pushing dev releases to pypi which means that pip install astroquery doesn t install the latest version of astroquery so what is the appropriate way to make this a fully automated release process can we just drop the dev from the version name does that break some of the astropy helpers machinery cc eteq bsipocz | 1 |
208,914 | 23,665,448,692 | IssuesEvent | 2022-08-26 20:19:37 | JohnDeere/work-tracker-examples | https://api.github.com/repos/JohnDeere/work-tracker-examples | closed | CVE-2020-36184 (High) detected in jackson-databind-2.9.10.3.jar - autoclosed | security vulnerability | ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /java-example/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.3/jackson-databind-2.9.10.3.jar,/com/fasterxml/jackson/core/jackson-databind/2.9.10.3/jackson-databind-2.9.10.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.3/jackson-databind-2.9.10.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/JohnDeere/work-tracker-examples/commit/7aa2fa9c80c3d14d7e62f0494ba7edaff8842068">7aa2fa9c80c3d14d7e62f0494ba7edaff8842068</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: 2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-36184 (High) detected in jackson-databind-2.9.10.3.jar - autoclosed - ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /java-example/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.3/jackson-databind-2.9.10.3.jar,/com/fasterxml/jackson/core/jackson-databind/2.9.10.3/jackson-databind-2.9.10.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.3/jackson-databind-2.9.10.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/JohnDeere/work-tracker-examples/commit/7aa2fa9c80c3d14d7e62f0494ba7edaff8842068">7aa2fa9c80c3d14d7e62f0494ba7edaff8842068</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: 2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file java example pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend | 0 |
6,731 | 6,575,758,651 | IssuesEvent | 2017-09-11 17:13:04 | empirical-org/Empirical-Core | https://api.github.com/repos/empirical-org/Empirical-Core | closed | Update rake:setup to use new sql structure | infrastructure | Schema no longer gets updated since we adopted gin indexes. This means devs have to run rake db:migrate after running the update task.
We should update the script to use the structure.sql file instead.
| 1.0 | Update rake:setup to use new sql structure - Schema no longer gets updated since we adopted gin indexes. This means devs have to run rake db:migrate after running the update task.
We should update the script to use the structure.sql file instead.
| infrastructure | update rake setup to use new sql structure schema no longer gets updated since we adopted gin indexes this means devs have to run rake db migrate after running the update task we should update the script to use the structure sql file instead | 1 |
28,685 | 23,443,120,561 | IssuesEvent | 2022-08-15 16:47:31 | JupiterBroadcasting/jupiterbroadcasting.com | https://api.github.com/repos/JupiterBroadcasting/jupiterbroadcasting.com | opened | Deployment smoke tests | enhancement infrastructure | We need some tests writing which will check that the hugo build was successful and the site deploys correctly. This will be used as part of a PR workflow before enabling merges to `main` once complete. For example if the tests fail for any reason, merging will be blocked until they pass.
I envisage this taking the form of checking the existence of critical files including configuration files and certain show pages. Then the site will be built (as part of an action most likely), spun up (again as part of the action) and then the smoketests will be performed to ensure the thing is running as expected. | 1.0 | Deployment smoke tests - We need some tests writing which will check that the hugo build was successful and the site deploys correctly. This will be used as part of a PR workflow before enabling merges to `main` once complete. For example if the tests fail for any reason, merging will be blocked until they pass.
I envisage this taking the form of checking the existence of critical files including configuration files and certain show pages. Then the site will be built (as part of an action most likely), spun up (again as part of the action) and then the smoketests will be performed to ensure the thing is running as expected. | infrastructure | deployment smoke tests we need some tests writing which will check that the hugo build was successful and the site deploys correctly this will be used as part of a pr workflow before enabling merges to main once complete for example if the tests fail for any reason merging will be blocked until they pass i envisage this taking the form of checking the existence of critical files including configuration files and certain show pages then the site will be built as part of an action most likely spun up again as part of the action and then the smoketests will be performed to ensure the thing is running as expected | 1 |
67,525 | 16,994,117,063 | IssuesEvent | 2021-07-01 02:39:46 | quarkusio/quarkus | https://api.github.com/repos/quarkusio/quarkus | closed | Make native image test use the test profile or not | area/build area/testing kind/bug | I know we had some conversations recently on the subject but I don't recall the conclusion.
It was related to the shift of hibernate's import file feature having different defaults in prod vs dev.
So let's capture what we really want here and see if we need code change and / or documentation changes before 1.0
Today @cescoffier and @agoncal got hit by IT run in prod and thus confusion. | 1.0 | Make native image test use the test profile or not - I know we had some conversations recently on the subject but I don't recall the conclusion.
It was related to the shift of hibernate's import file feature having different defaults in prod vs dev.
So let's capture what we really want here and see if we need code change and / or documentation changes before 1.0
Today @cescoffier and @agoncal got hit by IT run in prod and thus confusion. | non_infrastructure | make native image test use the test profile or not i know we had some conversations recently on the subject but i don t recall the conclusion it was related to the shift of hibernate s import file feature having different defaults in prod vs dev so let s capture what we really want here and see if we need code change and or documentation changes before today cescoffier and agoncal got hit by it run in prod and thus confusion | 0 |
26,751 | 20,648,562,204 | IssuesEvent | 2022-03-09 00:00:59 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Gazillion warnings from enabling building tests on Win8.1 x64 Debug | test-bug up-for-grabs area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate | I don't know why it's looking for UCRT libs in windows 10 directory paths for win8.1 build.
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd0E_r.csproj]
> hfa_nd0E_r -> c:\git\dotnet\coreclr\bin\tests\Windows_NT.x64.Debug\JIT\jit64\hfa\main\testE\hfa_nd0E_r\hfa_nd0E_r.exe
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x64' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x86' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x64' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_d.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x86' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_d.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_d.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x64' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd2E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x86' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd2E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd2E_r.csproj] | 1.0 | Gazillion warnings from enabling building tests on Win8.1 x64 Debug - I don't know why it's looking for UCRT libs in windows 10 directory paths for win8.1 build.
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd0E_r.csproj]
> hfa_nd0E_r -> c:\git\dotnet\coreclr\bin\tests\Windows_NT.x64.Debug\JIT\jit64\hfa\main\testE\hfa_nd0E_r\hfa_nd0E_r.exe
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x64' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x86' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x64' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_d.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x86' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_d.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd1E_d.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x64' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd2E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\lib\10.0.14393.0\ucrt\x86' specified in 'LIB en
> vironment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd2E_r.csproj]
> CSC : warning CS1668: Invalid search path 'C:\Program Files (x86)\Windows Kits\10\Lib\10.0.10240.0\um\x64' specified in 'LIB envi
> ronment variable' -- 'directory does not exist' [c:\git\dotnet\coreclr\tests\src\JIT\jit64\hfa\main\testE\hfa_nd2E_r.csproj] | infrastructure | gazillion warnings from enabling building tests on debug i don t know why it s looking for ucrt libs in windows directory paths for build csc warning invalid search path c program files windows kits lib um specified in lib envi ronment variable directory does not exist hfa r c git dotnet coreclr bin tests windows nt debug jit hfa main teste hfa r hfa r exe csc warning invalid search path c program files windows kits lib ucrt specified in lib en vironment variable directory does not exist csc warning invalid search path c program files windows kits lib ucrt specified in lib en vironment variable directory does not exist csc warning invalid search path c program files windows kits lib um specified in lib envi ronment variable directory does not exist csc warning invalid search path c program files windows kits lib ucrt specified in lib en vironment variable directory does not exist csc warning invalid search path c program files windows kits lib ucrt specified in lib en vironment variable directory does not exist csc warning invalid search path c program files windows kits lib um specified in lib envi ronment variable directory does not exist csc warning invalid search path c program files windows kits lib ucrt specified in lib en vironment variable directory does not exist csc warning invalid search path c program files windows kits lib ucrt specified in lib en vironment variable directory does not exist csc warning invalid search path c program files windows kits lib um specified in lib envi ronment variable directory does not exist | 1 |
246,320 | 20,833,936,205 | IssuesEvent | 2022-03-19 22:27:05 | commercialhaskell/stackage | https://api.github.com/repos/commercialhaskell/stackage | closed | fmt 0.6.3.0 | failure: test-suite | ```
Test suite failure for package fmt-0.6.3.0
doctests: exited with: ExitFailure 1
Full log available at /var/stackage/work/unpack-dir/.stack-work/logs/fmt-0.6.3.0-test.log
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt.hs:162: failure in expression `("Here are some words: "+|a|+", "+|b|+"\nAlso a number: "+|n|+"") :: String'
expected: "Here are some words: foo, bar\nAlso a number: 25"
but got:
^
<interactive>:41:25: error:
• Variable not in scope: (+|) :: t0 -> t1 -> t2
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:28: error:
• Variable not in scope: (|+) :: t2 -> t3 -> t4
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:34: error:
• Variable not in scope: (+|) :: t4 -> t5 -> t6
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:37: error:
• Variable not in scope: (|+) :: t6 -> t7 -> t8
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:58: error:
• Variable not in scope: (+|) :: t8 -> t9 -> t10
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:61: error:
• Variable not in scope: (|+) :: t10 -> t11 -> String
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt.hs:263: failure in expression `"Meet "+|name|+"!" :: String'
expected: "Meet Alice!"
but got:
^
<interactive>:55:8: error:
• Variable not in scope: (+|) :: t0 -> String -> t1
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:55:14: error:
• Variable not in scope: (|+) :: t1 -> t2 -> String
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt.hs:328: failure in expression `(""+|show foo|++|show bar|+"") == (""+||foo||++||bar||+"")'
<interactive>:78:2: error:
• No instance for (Show (Integer -> String -> Integer))
arising from a use of ‘Test.QuickCheck.Test.quickCheck’
(maybe you haven't applied a function to enough arguments?)
• In the expression:
Test.QuickCheck.Test.quickCheck
(doctest_prop ::
(Integer -> String -> Integer)
-> Integer
-> (Integer -> String -> Integer)
-> Integer
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer) -> Bool)
In an equation for ‘it’:
it
= (Test.QuickCheck.Test.quickCheck
(doctest_prop ::
(Integer -> String -> Integer)
-> Integer
-> (Integer -> String -> Integer)
-> Integer
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer) -> Bool))
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Time.hs:143: failure in expression `let t = read "2018-02-14 16:20:45.5 CST" :: ZonedTime'
expected:
but got:
^
<interactive>:84:45: error:
Not in scope: type constructor or class ‘ZonedTime’
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal.hs:51: failure in expression `import Fmt'
expected:
but got:
^
<interactive>:1:1: error:
attempting to use module ‘main:Fmt’ (lib/Fmt.hs) which is not loaded
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Generic.hs:39: failure in expression `data Foo = Foo { x :: Bool, y :: [Int] } deriving Generic'
expected:
but got:
^
<interactive>:104:51: error:
Not in scope: type constructor or class ‘Generic’
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Generic.hs:72: failure in expression `:{
data Bar = Bar { x :: Bool, y :: [Int] }
deriving stock Generic
deriving Buildable via GenericBuildable Bar
:}'
expected:
but got:
^
<interactive>:116:18: error:
Not in scope: type constructor or class ‘Generic’
<interactive>:117:12: error:
Not in scope: type constructor or class ‘Buildable’
<interactive>:117:26: error:
Not in scope: type constructor or class ‘GenericBuildable’
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Tuple.hs:20: failure in expression `import Fmt.Internal.Core'
expected:
but got:
^
<interactive>:1:1: error:
attempting to use module ‘main:Fmt.Internal.Core’ (lib/Fmt/Internal/Core.hs) which is not loaded
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Template.hs:32: failure in expression `format "{} + {} = {}" 2 2 4'
expected: 2 + 2 = 4
but got:
^
<interactive>:138:1: error:
Variable not in scope: format :: t0 -> t1 -> t2 -> t3 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:31: failure in expression `listF' octF [7,8,9,10]'
expected: "[7, 10, 11, 12]"
but got:
^
<interactive>:152:1: error:
Variable not in scope: listF' :: t0 -> [a0] -> t
<interactive>:152:8: error: Variable not in scope: octF
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:40: failure in expression `listF' binF [7,8,9,10]'
expected: "[111, 1000, 1001, 1010]"
but got:
^
<interactive>:160:1: error:
Variable not in scope: listF' :: t0 -> [a0] -> t
<interactive>:160:8: error: Variable not in scope: binF
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:49: failure in expression `baseF 3 10000'
expected: "111201101"
but got:
^
<interactive>:168:1: error:
Variable not in scope: baseF :: t0 -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:66: failure in expression `floatF 3.1415'
expected: "3.1415"
but got:
^
<interactive>:176:1: error:
Variable not in scope: floatF :: t0 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:88: failure in expression `listF' (exptF 5) [pi,0.1,10]'
expected: "[3.14159e0, 1.00000e-1, 1.00000e1]"
but got:
^
<interactive>:184:1: error:
Variable not in scope: listF' :: t1 -> [a0] -> t
<interactive>:184:9: error:
Variable not in scope: exptF :: t0 -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:97: failure in expression `listF' (fixedF 5) [pi,0.1,10]'
expected: "[3.14159, 0.10000, 10.00000]"
but got:
^
<interactive>:192:1: error:
Variable not in scope: listF' :: t1 -> [a0] -> t
<interactive>:192:9: error:
Variable not in scope: fixedF :: t0 -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:110: failure in expression `commaizeF 15830000'
expected: "15,830,000"
but got:
^
<interactive>:200:1: error:
Variable not in scope: commaizeF :: t0 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:119: failure in expression `ordinalF 15'
expected: "15th"
but got:
^
<interactive>:208:1: error:
Variable not in scope: ordinalF :: t0 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:48: failure in expression `fmt $ "This is a list:\n" <> indentF 4 (blockListF [1,2,3])'
expected: This is a list:
- 1
- 2
- 3
but got:
^
<interactive>:222:1: error:
• Variable not in scope: fmt :: t2 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:222:30: error:
Variable not in scope: indentF :: t0 -> t1 -> t2
<interactive>:222:41: error:
Variable not in scope: blockListF :: [a0] -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:77: failure in expression `fmt $ nameF "clients" $ blockListF ["Alice", "Bob", "Zalgo"]'
expected: clients:
- Alice
- Bob
- Zalgo
but got:
^
<interactive>:230:1: error:
• Variable not in scope: fmt :: t2 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:230:7: error:
Variable not in scope: nameF :: t0 -> t1 -> t2
<interactive>:230:25: error:
Variable not in scope: blockListF :: [a0] -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:92: failure in expression `fmt $ unwordsF ["hello", "world"]'
expected: hello world
but got:
^
<interactive>:238:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:238:7: error:
• Variable not in scope: unwordsF :: [a0] -> t0
• Perhaps you meant ‘unwords’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:107: failure in expression `fmt $ unlinesF ["hello", "world"]'
expected: hello
world
but got:
^
<interactive>:246:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:246:7: error:
• Variable not in scope: unlinesF :: [a0] -> t0
• Perhaps you meant ‘unlines’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:125: failure in expression `listF ["hello", "world"]'
expected: "[hello, world]"
but got:
^
<interactive>:254:1: error:
Variable not in scope: listF :: [a0] -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:139: failure in expression `listF' hexF [1234, 5678]'
expected: "[4d2, 162e]"
but got:
^
<interactive>:262:1: error:
Variable not in scope: listF' :: t0 -> [a0] -> t
<interactive>:262:8: error: Variable not in scope: hexF
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:169: failure in expression `fmt $ blockListF [1,2,3]'
expected: - 1
- 2
- 3
but got:
^
<interactive>:270:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:270:7: error:
Variable not in scope: blockListF :: [a0] -> t0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:211: failure in expression `fmt $ jsonListF [1,2,3]'
expected: [
1
, 2
, 3
]
but got:
^
<interactive>:278:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:278:7: error:
Variable not in scope: jsonListF :: [a0] -> t0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:262: failure in expression `mapF [("a", 1), ("b", 4)]'
expected: "{a: 1, b: 4}"
but got:
^
<interactive>:286:1: error:
• Variable not in scope: mapF :: [(a0, b0)] -> t
• Perhaps you meant one of these:
‘map’ (imported from Prelude), ‘mapM’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:284: failure in expression `fmt $ blockMapF [("Odds", blockListF [1,3]), ("Evens", blockListF [2,4])]'
expected: Odds:
- 1
- 3
Evens:
- 2
- 4
but got:
^
<interactive>:294:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:294:7: error:
Variable not in scope: blockMapF :: [(a2, b0)] -> t0
<interactive>:294:27: error:
Variable not in scope: blockListF :: [a1] -> b0
<interactive>:294:56: error:
Variable not in scope: blockListF :: [a0] -> b0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:310: failure in expression `fmt $ jsonMapF [("Odds", jsonListF [1,3]), ("Evens", jsonListF [2,4])]'
expected: {
Odds:
[
1
, 3
]
, Evens:
[
2
, 4
]
}
but got:
^
<interactive>:302:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:302:7: error:
Variable not in scope: jsonMapF :: [(a2, b0)] -> t0
<interactive>:302:26: error:
Variable not in scope: jsonListF :: [a1] -> b0
<interactive>:302:54: error:
Variable not in scope: jsonListF :: [a0] -> b0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:359: failure in expression `build (Nothing :: Maybe Int)'
expected: ""
but got:
^
<interactive>:310:1: error:
Variable not in scope: build :: Maybe Int -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:377: failure in expression `eitherF (Right 1 :: Either Bool Int)'
expected: "<Right: 1>"
but got:
^
<interactive>:318:1: error:
• Variable not in scope: eitherF :: Either Bool Int -> t
• Perhaps you meant ‘either’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:391: failure in expression `prefixF 3 "hello"'
expected: "hel"
but got:
^
<interactive>:326:1: error:
Variable not in scope: prefixF :: t0 -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:401: failure in expression `suffixF 3 "hello"'
expected: "llo"
but got:
^
<interactive>:334:1: error:
Variable not in scope: suffixF :: t0 -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:415: failure in expression `padLeftF 5 '0' 12'
expected: "00012"
but got:
^
<interactive>:342:1: error:
Variable not in scope: padLeftF :: t0 -> Char -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:428: failure in expression `padRightF 5 ' ' "foo"'
expected: "foo "
but got:
^
<interactive>:350:1: error:
Variable not in scope: padRightF :: t0 -> Char -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:441: failure in expression `padBothF 5 '=' "foo"'
expected: "=foo="
but got:
^
<interactive>:358:1: error:
Variable not in scope: padBothF :: t0 -> Char -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Core.hs:125: failure in expression `fmt "Hello world!\n"'
expected: Hello world!
but got:
^
<interactive>:364:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
Examples: 147 Tried: 46 Errors: 0 Failures: 36
```
cc @cdornan | 1.0 | fmt 0.6.3.0 - ```
Test suite failure for package fmt-0.6.3.0
doctests: exited with: ExitFailure 1
Full log available at /var/stackage/work/unpack-dir/.stack-work/logs/fmt-0.6.3.0-test.log
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt.hs:162: failure in expression `("Here are some words: "+|a|+", "+|b|+"\nAlso a number: "+|n|+"") :: String'
expected: "Here are some words: foo, bar\nAlso a number: 25"
but got:
^
<interactive>:41:25: error:
• Variable not in scope: (+|) :: t0 -> t1 -> t2
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:28: error:
• Variable not in scope: (|+) :: t2 -> t3 -> t4
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:34: error:
• Variable not in scope: (+|) :: t4 -> t5 -> t6
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:37: error:
• Variable not in scope: (|+) :: t6 -> t7 -> t8
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:58: error:
• Variable not in scope: (+|) :: t8 -> t9 -> t10
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:41:61: error:
• Variable not in scope: (|+) :: t10 -> t11 -> String
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt.hs:263: failure in expression `"Meet "+|name|+"!" :: String'
expected: "Meet Alice!"
but got:
^
<interactive>:55:8: error:
• Variable not in scope: (+|) :: t0 -> String -> t1
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
<interactive>:55:14: error:
• Variable not in scope: (|+) :: t1 -> t2 -> String
• Perhaps you meant one of these:
‘+’ (imported from Prelude), ‘++’ (imported from Prelude),
‘||’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt.hs:328: failure in expression `(""+|show foo|++|show bar|+"") == (""+||foo||++||bar||+"")'
<interactive>:78:2: error:
• No instance for (Show (Integer -> String -> Integer))
arising from a use of ‘Test.QuickCheck.Test.quickCheck’
(maybe you haven't applied a function to enough arguments?)
• In the expression:
Test.QuickCheck.Test.quickCheck
(doctest_prop ::
(Integer -> String -> Integer)
-> Integer
-> (Integer -> String -> Integer)
-> Integer
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer) -> Bool)
In an equation for ‘it’:
it
= (Test.QuickCheck.Test.quickCheck
(doctest_prop ::
(Integer -> String -> Integer)
-> Integer
-> (Integer -> String -> Integer)
-> Integer
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer)
-> (Integer -> Integer -> Integer) -> Bool))
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Time.hs:143: failure in expression `let t = read "2018-02-14 16:20:45.5 CST" :: ZonedTime'
expected:
but got:
^
<interactive>:84:45: error:
Not in scope: type constructor or class ‘ZonedTime’
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal.hs:51: failure in expression `import Fmt'
expected:
but got:
^
<interactive>:1:1: error:
attempting to use module ‘main:Fmt’ (lib/Fmt.hs) which is not loaded
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Generic.hs:39: failure in expression `data Foo = Foo { x :: Bool, y :: [Int] } deriving Generic'
expected:
but got:
^
<interactive>:104:51: error:
Not in scope: type constructor or class ‘Generic’
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Generic.hs:72: failure in expression `:{
data Bar = Bar { x :: Bool, y :: [Int] }
deriving stock Generic
deriving Buildable via GenericBuildable Bar
:}'
expected:
but got:
^
<interactive>:116:18: error:
Not in scope: type constructor or class ‘Generic’
<interactive>:117:12: error:
Not in scope: type constructor or class ‘Buildable’
<interactive>:117:26: error:
Not in scope: type constructor or class ‘GenericBuildable’
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Tuple.hs:20: failure in expression `import Fmt.Internal.Core'
expected:
but got:
^
<interactive>:1:1: error:
attempting to use module ‘main:Fmt.Internal.Core’ (lib/Fmt/Internal/Core.hs) which is not loaded
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Template.hs:32: failure in expression `format "{} + {} = {}" 2 2 4'
expected: 2 + 2 = 4
but got:
^
<interactive>:138:1: error:
Variable not in scope: format :: t0 -> t1 -> t2 -> t3 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:31: failure in expression `listF' octF [7,8,9,10]'
expected: "[7, 10, 11, 12]"
but got:
^
<interactive>:152:1: error:
Variable not in scope: listF' :: t0 -> [a0] -> t
<interactive>:152:8: error: Variable not in scope: octF
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:40: failure in expression `listF' binF [7,8,9,10]'
expected: "[111, 1000, 1001, 1010]"
but got:
^
<interactive>:160:1: error:
Variable not in scope: listF' :: t0 -> [a0] -> t
<interactive>:160:8: error: Variable not in scope: binF
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:49: failure in expression `baseF 3 10000'
expected: "111201101"
but got:
^
<interactive>:168:1: error:
Variable not in scope: baseF :: t0 -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:66: failure in expression `floatF 3.1415'
expected: "3.1415"
but got:
^
<interactive>:176:1: error:
Variable not in scope: floatF :: t0 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:88: failure in expression `listF' (exptF 5) [pi,0.1,10]'
expected: "[3.14159e0, 1.00000e-1, 1.00000e1]"
but got:
^
<interactive>:184:1: error:
Variable not in scope: listF' :: t1 -> [a0] -> t
<interactive>:184:9: error:
Variable not in scope: exptF :: t0 -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:97: failure in expression `listF' (fixedF 5) [pi,0.1,10]'
expected: "[3.14159, 0.10000, 10.00000]"
but got:
^
<interactive>:192:1: error:
Variable not in scope: listF' :: t1 -> [a0] -> t
<interactive>:192:9: error:
Variable not in scope: fixedF :: t0 -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:110: failure in expression `commaizeF 15830000'
expected: "15,830,000"
but got:
^
<interactive>:200:1: error:
Variable not in scope: commaizeF :: t0 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Numeric.hs:119: failure in expression `ordinalF 15'
expected: "15th"
but got:
^
<interactive>:208:1: error:
Variable not in scope: ordinalF :: t0 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:48: failure in expression `fmt $ "This is a list:\n" <> indentF 4 (blockListF [1,2,3])'
expected: This is a list:
- 1
- 2
- 3
but got:
^
<interactive>:222:1: error:
• Variable not in scope: fmt :: t2 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:222:30: error:
Variable not in scope: indentF :: t0 -> t1 -> t2
<interactive>:222:41: error:
Variable not in scope: blockListF :: [a0] -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:77: failure in expression `fmt $ nameF "clients" $ blockListF ["Alice", "Bob", "Zalgo"]'
expected: clients:
- Alice
- Bob
- Zalgo
but got:
^
<interactive>:230:1: error:
• Variable not in scope: fmt :: t2 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:230:7: error:
Variable not in scope: nameF :: t0 -> t1 -> t2
<interactive>:230:25: error:
Variable not in scope: blockListF :: [a0] -> t1
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:92: failure in expression `fmt $ unwordsF ["hello", "world"]'
expected: hello world
but got:
^
<interactive>:238:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:238:7: error:
• Variable not in scope: unwordsF :: [a0] -> t0
• Perhaps you meant ‘unwords’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:107: failure in expression `fmt $ unlinesF ["hello", "world"]'
expected: hello
world
but got:
^
<interactive>:246:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:246:7: error:
• Variable not in scope: unlinesF :: [a0] -> t0
• Perhaps you meant ‘unlines’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:125: failure in expression `listF ["hello", "world"]'
expected: "[hello, world]"
but got:
^
<interactive>:254:1: error:
Variable not in scope: listF :: [a0] -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:139: failure in expression `listF' hexF [1234, 5678]'
expected: "[4d2, 162e]"
but got:
^
<interactive>:262:1: error:
Variable not in scope: listF' :: t0 -> [a0] -> t
<interactive>:262:8: error: Variable not in scope: hexF
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:169: failure in expression `fmt $ blockListF [1,2,3]'
expected: - 1
- 2
- 3
but got:
^
<interactive>:270:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:270:7: error:
Variable not in scope: blockListF :: [a0] -> t0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:211: failure in expression `fmt $ jsonListF [1,2,3]'
expected: [
1
, 2
, 3
]
but got:
^
<interactive>:278:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:278:7: error:
Variable not in scope: jsonListF :: [a0] -> t0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:262: failure in expression `mapF [("a", 1), ("b", 4)]'
expected: "{a: 1, b: 4}"
but got:
^
<interactive>:286:1: error:
• Variable not in scope: mapF :: [(a0, b0)] -> t
• Perhaps you meant one of these:
‘map’ (imported from Prelude), ‘mapM’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:284: failure in expression `fmt $ blockMapF [("Odds", blockListF [1,3]), ("Evens", blockListF [2,4])]'
expected: Odds:
- 1
- 3
Evens:
- 2
- 4
but got:
^
<interactive>:294:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:294:7: error:
Variable not in scope: blockMapF :: [(a2, b0)] -> t0
<interactive>:294:27: error:
Variable not in scope: blockListF :: [a1] -> b0
<interactive>:294:56: error:
Variable not in scope: blockListF :: [a0] -> b0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:310: failure in expression `fmt $ jsonMapF [("Odds", jsonListF [1,3]), ("Evens", jsonListF [2,4])]'
expected: {
Odds:
[
1
, 3
]
, Evens:
[
2
, 4
]
}
but got:
^
<interactive>:302:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
<interactive>:302:7: error:
Variable not in scope: jsonMapF :: [(a2, b0)] -> t0
<interactive>:302:26: error:
Variable not in scope: jsonListF :: [a1] -> b0
<interactive>:302:54: error:
Variable not in scope: jsonListF :: [a0] -> b0
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:359: failure in expression `build (Nothing :: Maybe Int)'
expected: ""
but got:
^
<interactive>:310:1: error:
Variable not in scope: build :: Maybe Int -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:377: failure in expression `eitherF (Right 1 :: Either Bool Int)'
expected: "<Right: 1>"
but got:
^
<interactive>:318:1: error:
• Variable not in scope: eitherF :: Either Bool Int -> t
• Perhaps you meant ‘either’ (imported from Prelude)
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:391: failure in expression `prefixF 3 "hello"'
expected: "hel"
but got:
^
<interactive>:326:1: error:
Variable not in scope: prefixF :: t0 -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:401: failure in expression `suffixF 3 "hello"'
expected: "llo"
but got:
^
<interactive>:334:1: error:
Variable not in scope: suffixF :: t0 -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:415: failure in expression `padLeftF 5 '0' 12'
expected: "00012"
but got:
^
<interactive>:342:1: error:
Variable not in scope: padLeftF :: t0 -> Char -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:428: failure in expression `padRightF 5 ' ' "foo"'
expected: "foo "
but got:
^
<interactive>:350:1: error:
Variable not in scope: padRightF :: t0 -> Char -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Formatters.hs:441: failure in expression `padBothF 5 '=' "foo"'
expected: "=foo="
but got:
^
<interactive>:358:1: error:
Variable not in scope: padBothF :: t0 -> Char -> t1 -> t
/var/stackage/work/unpack-dir/unpacked/fmt-0.6.3.0-19b0d53555f2789b99898d17118a3dab3cc8d7bb26ec7c9d2553331f528ef58e/lib/Fmt/Internal/Core.hs:125: failure in expression `fmt "Hello world!\n"'
expected: Hello world!
but got:
^
<interactive>:364:1: error:
• Variable not in scope: fmt :: t0 -> t
• Perhaps you meant ‘fst’ (imported from Prelude)
Examples: 147 Tried: 46 Errors: 0 Failures: 36
```
cc @cdornan | non_infrastructure | fmt test suite failure for package fmt doctests exited with exitfailure full log available at var stackage work unpack dir stack work logs fmt test log var stackage work unpack dir unpacked fmt lib fmt hs failure in expression here are some words a b nalso a number n string expected here are some words foo bar nalso a number but got error • variable not in scope • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude error • variable not in scope • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude error • variable not in scope • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude error • variable not in scope • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude error • variable not in scope • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude error • variable not in scope string • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude var stackage work unpack dir unpacked fmt lib fmt hs failure in expression meet name string expected meet alice but got error • variable not in scope string • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude error • variable not in scope string • perhaps you meant one of these ‘ ’ imported from prelude ‘ ’ imported from prelude ‘ ’ imported from prelude var stackage work unpack dir unpacked fmt lib fmt hs failure in expression show foo show bar foo bar error • no instance for show integer string integer arising from a use of ‘test quickcheck test quickcheck’ maybe you haven t applied a function to enough arguments • in the expression test quickcheck test quickcheck doctest prop integer string integer integer integer string integer integer integer integer integer integer integer integer integer integer integer integer integer integer bool in an equation for ‘it’ it test quickcheck test quickcheck doctest prop integer string integer integer integer string integer integer integer integer integer integer integer integer integer integer integer integer integer integer bool var stackage work unpack dir unpacked fmt lib fmt time hs failure in expression let t read cst zonedtime expected but got error not in scope type constructor or class ‘zonedtime’ var stackage work unpack dir unpacked fmt lib fmt internal hs failure in expression import fmt expected but got error attempting to use module ‘main fmt’ lib fmt hs which is not loaded var stackage work unpack dir unpacked fmt lib fmt internal generic hs failure in expression data foo foo x bool y deriving generic expected but got error not in scope type constructor or class ‘generic’ var stackage work unpack dir unpacked fmt lib fmt internal generic hs failure in expression data bar bar x bool y deriving stock generic deriving buildable via genericbuildable bar expected but got error not in scope type constructor or class ‘generic’ error not in scope type constructor or class ‘buildable’ error not in scope type constructor or class ‘genericbuildable’ var stackage work unpack dir unpacked fmt lib fmt internal tuple hs failure in expression import fmt internal core expected but got error attempting to use module ‘main fmt internal core’ lib fmt internal core hs which is not loaded var stackage work unpack dir unpacked fmt lib fmt internal template hs failure in expression format expected but got error variable not in scope format t var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression listf octf expected but got error variable not in scope listf t error variable not in scope octf var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression listf binf expected but got error variable not in scope listf t error variable not in scope binf var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression basef expected but got error variable not in scope basef t var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression floatf expected but got error variable not in scope floatf t var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression listf exptf expected but got error variable not in scope listf t error variable not in scope exptf var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression listf fixedf expected but got error variable not in scope listf t error variable not in scope fixedf var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression commaizef expected but got error variable not in scope commaizef t var stackage work unpack dir unpacked fmt lib fmt internal numeric hs failure in expression ordinalf expected but got error variable not in scope ordinalf t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt this is a list n indentf blocklistf expected this is a list but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error variable not in scope indentf error variable not in scope blocklistf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt namef clients blocklistf expected clients alice bob zalgo but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error variable not in scope namef error variable not in scope blocklistf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt unwordsf expected hello world but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error • variable not in scope unwordsf • perhaps you meant ‘unwords’ imported from prelude var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt unlinesf expected hello world but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error • variable not in scope unlinesf • perhaps you meant ‘unlines’ imported from prelude var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression listf expected but got error variable not in scope listf t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression listf hexf expected but got error variable not in scope listf t error variable not in scope hexf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt blocklistf expected but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error variable not in scope blocklistf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt jsonlistf expected but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error variable not in scope jsonlistf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression mapf expected a b but got error • variable not in scope mapf t • perhaps you meant one of these ‘map’ imported from prelude ‘mapm’ imported from prelude var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt blockmapf evens blocklistf expected odds evens but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error variable not in scope blockmapf error variable not in scope blocklistf error variable not in scope blocklistf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression fmt jsonmapf evens jsonlistf expected odds evens but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude error variable not in scope jsonmapf error variable not in scope jsonlistf error variable not in scope jsonlistf var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression build nothing maybe int expected but got error variable not in scope build maybe int t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression eitherf right either bool int expected but got error • variable not in scope eitherf either bool int t • perhaps you meant ‘either’ imported from prelude var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression prefixf hello expected hel but got error variable not in scope prefixf t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression suffixf hello expected llo but got error variable not in scope suffixf t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression padleftf expected but got error variable not in scope padleftf char t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression padrightf foo expected foo but got error variable not in scope padrightf char t var stackage work unpack dir unpacked fmt lib fmt internal formatters hs failure in expression padbothf foo expected foo but got error variable not in scope padbothf char t var stackage work unpack dir unpacked fmt lib fmt internal core hs failure in expression fmt hello world n expected hello world but got error • variable not in scope fmt t • perhaps you meant ‘fst’ imported from prelude examples tried errors failures cc cdornan | 0 |
25,247 | 18,298,713,114 | IssuesEvent | 2021-10-05 23:29:38 | IntelPython/dpctl | https://api.github.com/repos/IntelPython/dpctl | closed | Fix CI job `Generate coverage data / Generate coverage and push to Coveralls.io` | infrastructure | I have made job "Generate coverage data / Generate coverage and push to Coveralls.io" as not required because it fails on master too.
_Originally posted by @PokhodenkoSA in https://github.com/IntelPython/dpctl/issues/599#issuecomment-931142866_ | 1.0 | Fix CI job `Generate coverage data / Generate coverage and push to Coveralls.io` - I have made job "Generate coverage data / Generate coverage and push to Coveralls.io" as not required because it fails on master too.
_Originally posted by @PokhodenkoSA in https://github.com/IntelPython/dpctl/issues/599#issuecomment-931142866_ | infrastructure | fix ci job generate coverage data generate coverage and push to coveralls io i have made job generate coverage data generate coverage and push to coveralls io as not required because it fails on master too originally posted by pokhodenkosa in | 1 |
74,989 | 20,596,348,189 | IssuesEvent | 2022-03-05 15:07:50 | libical/libical | https://api.github.com/repos/libical/libical | closed | CMAKE_SOURCE_DIR and CMAKE_BINARY_DIR replace with PROJECT_SOURCE_DIR and PROJECT_BINARY_DIR | buildsystem | replace CMAKE_SOURCE_DIR and CMAKE_BINARY_DIR with PROJECT_SOURCE_DIR and PROJECT_BINARY_DIR
allows easier use of libical as a submodule | 1.0 | CMAKE_SOURCE_DIR and CMAKE_BINARY_DIR replace with PROJECT_SOURCE_DIR and PROJECT_BINARY_DIR - replace CMAKE_SOURCE_DIR and CMAKE_BINARY_DIR with PROJECT_SOURCE_DIR and PROJECT_BINARY_DIR
allows easier use of libical as a submodule | non_infrastructure | cmake source dir and cmake binary dir replace with project source dir and project binary dir replace cmake source dir and cmake binary dir with project source dir and project binary dir allows easier use of libical as a submodule | 0 |
160,538 | 6,099,354,261 | IssuesEvent | 2017-06-20 09:50:22 | jiscdev/xapi | https://api.github.com/repos/jiscdev/xapi | closed | subType | awaiting feedback for weekly call priority question | I've been looking at generic statement templates, and I don't quite yet understand subType.
There are various indications that subType is used as given in the Vocabulary document. However, that simply defines it as "Identifies the sub genre of any application" with only 1 example (and that example is defined in tincan as a platform). Do we have a list of subTypes somewhere?
Alan | 1.0 | subType - I've been looking at generic statement templates, and I don't quite yet understand subType.
There are various indications that subType is used as given in the Vocabulary document. However, that simply defines it as "Identifies the sub genre of any application" with only 1 example (and that example is defined in tincan as a platform). Do we have a list of subTypes somewhere?
Alan | non_infrastructure | subtype i ve been looking at generic statement templates and i don t quite yet understand subtype there are various indications that subtype is used as given in the vocabulary document however that simply defines it as identifies the sub genre of any application with only example and that example is defined in tincan as a platform do we have a list of subtypes somewhere alan | 0 |
16,349 | 11,941,141,695 | IssuesEvent | 2020-04-02 17:55:05 | cockroachdb/docs | https://api.github.com/repos/cockroachdb/docs | closed | Smoothen automatable tasks | A-docs-infrastructure T-enhancement | - [ ] markup around diagrams to avoid markdown from going haywire
- [ ] js for copy-paste before examples
- [ ] clean up extant examples before doing so
- [x] div toc at the beginning
- [ ] automate the layout and indentation of SQL examples
| 1.0 | Smoothen automatable tasks - - [ ] markup around diagrams to avoid markdown from going haywire
- [ ] js for copy-paste before examples
- [ ] clean up extant examples before doing so
- [x] div toc at the beginning
- [ ] automate the layout and indentation of SQL examples
| infrastructure | smoothen automatable tasks markup around diagrams to avoid markdown from going haywire js for copy paste before examples clean up extant examples before doing so div toc at the beginning automate the layout and indentation of sql examples | 1 |
76,488 | 14,629,567,678 | IssuesEvent | 2020-12-23 16:04:37 | Componolit/systematization-binary-vulnerabilities | https://api.github.com/repos/Componolit/systematization-binary-vulnerabilities | opened | CVE-2020-25109: AMNESIA:33 | DNS Denial of Service Remote Code Execution | # Description
> The number of DNS queries/responses (set in DNS header) is not checked against the data present
# Root cause
> An issue was discovered in the DNS implementation in Ethernut in Nut/OS 5.1. The number of DNS queries/responses (set in a DNS header) is not checked against the data present. This may lead to successful Denial-of-Service, and possibly Remote Code Execution.
# Links
- [CVE Entry](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-25109)
- [Report](https://www.forescout.com/company/resources/amnesia33-how-tcp-ip-stacks-breed-critical-vulnerabilities-in-iot-ot-and-it-devices/)
| 1.0 | CVE-2020-25109: AMNESIA:33 - # Description
> The number of DNS queries/responses (set in DNS header) is not checked against the data present
# Root cause
> An issue was discovered in the DNS implementation in Ethernut in Nut/OS 5.1. The number of DNS queries/responses (set in a DNS header) is not checked against the data present. This may lead to successful Denial-of-Service, and possibly Remote Code Execution.
# Links
- [CVE Entry](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-25109)
- [Report](https://www.forescout.com/company/resources/amnesia33-how-tcp-ip-stacks-breed-critical-vulnerabilities-in-iot-ot-and-it-devices/)
| non_infrastructure | cve amnesia description the number of dns queries responses set in dns header is not checked against the data present root cause an issue was discovered in the dns implementation in ethernut in nut os the number of dns queries responses set in a dns header is not checked against the data present this may lead to successful denial of service and possibly remote code execution links | 0 |
34,958 | 30,624,347,590 | IssuesEvent | 2023-07-24 10:23:44 | microsoftgraph/microsoft-graph-devx-api | https://api.github.com/repos/microsoftgraph/microsoft-graph-devx-api | opened | Improve reliability metrics | type: epic type: infrastructure | - [ ] Improve 4XX error rate by:
- improving server side url validation using Kibali
- improve client side validation
- solve the problem of openapi paths overgeneration
- [ ] improve 5XX error rate which involves instigating failing OpenAPI snippet requests especially for Powershell. Target from 0.5% to 0% | 1.0 | Improve reliability metrics - - [ ] Improve 4XX error rate by:
- improving server side url validation using Kibali
- improve client side validation
- solve the problem of openapi paths overgeneration
- [ ] improve 5XX error rate which involves instigating failing OpenAPI snippet requests especially for Powershell. Target from 0.5% to 0% | infrastructure | improve reliability metrics improve error rate by improving server side url validation using kibali improve client side validation solve the problem of openapi paths overgeneration improve error rate which involves instigating failing openapi snippet requests especially for powershell target from to | 1 |
452,399 | 13,049,780,078 | IssuesEvent | 2020-07-29 14:33:56 | ChainSafe/gossamer | https://api.github.com/repos/ChainSafe/gossamer | closed | dot/network: implement max peers on network layer | Priority: 2 - High cmd network | <!---
PLEASE READ CAREFULLY
-->
## Expected Behavior
<!---
If you're describing a bug, tell us what should happen.
If you're suggesting a change/improvement, tell us how it should work.
-->
- there should be a config option for maximum network peers
- by default, there *ideally* be 50 peers (+10 validator peers, if validator node)
- 25 slots are for established connections, 25 for new incoming connections
## Current Behavior
<!---
If describing a bug, tell us what happens instead of the expected behavior.
If suggesting a change or an improvement, explain the difference between your
suggestion and current behavior.
-->
- no peer limit configured
- NOTE: the node runs out of memory with too many peers, so 50 peers may not be possible at this point (see #1028)
## Checklist
<!---
Each empty square brackets below is a checkbox. Replace [ ] with [x] to check
the box after completing the task.
--->
- [x] I have read [CODE_OF_CONDUCT](https://github.com/ChainSafe/gossamer/blob/development/.github/CODE_OF_CONDUCT.md) and [CONTRIBUTING](https://github.com/ChainSafe/gossamer/blob/development/.github/CONTRIBUTING.md)
- [x] I have provided as much information as possible and necessary
- [x] I am planning to submit a pull request to fix this issue myself
| 1.0 | dot/network: implement max peers on network layer - <!---
PLEASE READ CAREFULLY
-->
## Expected Behavior
<!---
If you're describing a bug, tell us what should happen.
If you're suggesting a change/improvement, tell us how it should work.
-->
- there should be a config option for maximum network peers
- by default, there *ideally* be 50 peers (+10 validator peers, if validator node)
- 25 slots are for established connections, 25 for new incoming connections
## Current Behavior
<!---
If describing a bug, tell us what happens instead of the expected behavior.
If suggesting a change or an improvement, explain the difference between your
suggestion and current behavior.
-->
- no peer limit configured
- NOTE: the node runs out of memory with too many peers, so 50 peers may not be possible at this point (see #1028)
## Checklist
<!---
Each empty square brackets below is a checkbox. Replace [ ] with [x] to check
the box after completing the task.
--->
- [x] I have read [CODE_OF_CONDUCT](https://github.com/ChainSafe/gossamer/blob/development/.github/CODE_OF_CONDUCT.md) and [CONTRIBUTING](https://github.com/ChainSafe/gossamer/blob/development/.github/CONTRIBUTING.md)
- [x] I have provided as much information as possible and necessary
- [x] I am planning to submit a pull request to fix this issue myself
| non_infrastructure | dot network implement max peers on network layer please read carefully expected behavior if you re describing a bug tell us what should happen if you re suggesting a change improvement tell us how it should work there should be a config option for maximum network peers by default there ideally be peers validator peers if validator node slots are for established connections for new incoming connections current behavior if describing a bug tell us what happens instead of the expected behavior if suggesting a change or an improvement explain the difference between your suggestion and current behavior no peer limit configured note the node runs out of memory with too many peers so peers may not be possible at this point see checklist each empty square brackets below is a checkbox replace with to check the box after completing the task i have read and i have provided as much information as possible and necessary i am planning to submit a pull request to fix this issue myself | 0 |
253,882 | 21,711,499,196 | IssuesEvent | 2022-05-10 14:13:30 | yildizmus/bounSWE573-2022 | https://api.github.com/repos/yildizmus/bounSWE573-2022 | opened | Getting started with unit testing | testing | Getting test report using coverage, performing first test in test.py in local repo, and automating test operations using github action | 1.0 | Getting started with unit testing - Getting test report using coverage, performing first test in test.py in local repo, and automating test operations using github action | non_infrastructure | getting started with unit testing getting test report using coverage performing first test in test py in local repo and automating test operations using github action | 0 |
6,406 | 3,020,873,770 | IssuesEvent | 2015-07-31 11:02:07 | pocoproject/poco | https://api.github.com/repos/pocoproject/poco | closed | Use Semantic Versioning for future releases | documentation | Starting with release 1.6.0, we'll use Semantic Versioning 2.0 (http://semver.org/spec/v2.0.0.html) for all subsequent POCO releases. This means the next release after 1.6.0 just fixing bugs will be 1.6.1, not 1.6.0p1. The next "feature" release after 1.6.0 will be 1.7.0. | 1.0 | Use Semantic Versioning for future releases - Starting with release 1.6.0, we'll use Semantic Versioning 2.0 (http://semver.org/spec/v2.0.0.html) for all subsequent POCO releases. This means the next release after 1.6.0 just fixing bugs will be 1.6.1, not 1.6.0p1. The next "feature" release after 1.6.0 will be 1.7.0. | non_infrastructure | use semantic versioning for future releases starting with release we ll use semantic versioning for all subsequent poco releases this means the next release after just fixing bugs will be not the next feature release after will be | 0 |
36,436 | 17,690,231,129 | IssuesEvent | 2021-08-24 09:02:33 | openvinotoolkit/openvino | https://api.github.com/repos/openvinotoolkit/openvino | closed | [Bug]Please Advise how to use the Accuracy-Check to Get the Average Precision Data | performance support_request | Hi,
Currently I am using the following OpenVINO models to get the AP data.
Person-Detection-Retail-00002 / Person-Detection-Retail-00013 / Person-Detection-0106.
Within three models I have used, I have found that 0106 model is the best based on the accuracy calculation scripts I used.
Now I need to use Intel Accuracy-Check to get the AP. But I can never get a good data. Hope you can help me on this!
Dataset Name: COCO 2017 CrowdHuman My Own Dataset
-----------------------------------------------------------------------------------------------------------------------------------------------------
Person-Detection-Retail-00002 AP? AP? AP?
Person-Detection-Retail-00013 AP? AP? AP?
Person-Detection-0106. AP? AP? AP?
Here is the dataset annotation format:
Dataset Name: COCO 2017 CrowdHuman My Own Dataset
=========================================================================
KITTI TXT .odgt KITTI TXT
Can you please let me know if Accuracy-Check can handle above annotation format since I have the big problem to get the accuracy check to run?
Thanks a lot
Jimmy
| True | [Bug]Please Advise how to use the Accuracy-Check to Get the Average Precision Data - Hi,
Currently I am using the following OpenVINO models to get the AP data.
Person-Detection-Retail-00002 / Person-Detection-Retail-00013 / Person-Detection-0106.
Within three models I have used, I have found that 0106 model is the best based on the accuracy calculation scripts I used.
Now I need to use Intel Accuracy-Check to get the AP. But I can never get a good data. Hope you can help me on this!
Dataset Name: COCO 2017 CrowdHuman My Own Dataset
-----------------------------------------------------------------------------------------------------------------------------------------------------
Person-Detection-Retail-00002 AP? AP? AP?
Person-Detection-Retail-00013 AP? AP? AP?
Person-Detection-0106. AP? AP? AP?
Here is the dataset annotation format:
Dataset Name: COCO 2017 CrowdHuman My Own Dataset
=========================================================================
KITTI TXT .odgt KITTI TXT
Can you please let me know if Accuracy-Check can handle above annotation format since I have the big problem to get the accuracy check to run?
Thanks a lot
Jimmy
| non_infrastructure | please advise how to use the accuracy check to get the average precision data hi currently i am using the following openvino models to get the ap data person detection retail person detection retail person detection within three models i have used i have found that model is the best based on the accuracy calculation scripts i used now i need to use intel accuracy check to get the ap but i can never get a good data hope you can help me on this dataset name coco crowdhuman my own dataset person detection retail ap ap ap person detection retail ap ap ap person detection ap ap ap here is the dataset annotation format dataset name coco crowdhuman my own dataset kitti txt odgt kitti txt can you please let me know if accuracy check can handle above annotation format since i have the big problem to get the accuracy check to run thanks a lot jimmy | 0 |
301,146 | 26,020,576,775 | IssuesEvent | 2022-12-21 12:17:20 | stianst/playground | https://api.github.com/repos/stianst/playground | closed | Flaky test: playground.stianst.github.io.MyFlakyTest#flaky2 | flaky-test kind/bug area/ci | ## playground.stianst.github.io.MyFlakyTest#flaky2
[Run (pull_request)](https://github.com/stianst/playground/actions/runs/3748377827) / [Pull Request #60](https://github.com/stianst/playground/pull/60)
### Errors
```
java.lang.AssertionError
at playground.stianst.github.io.MyFlakyTest.flaky2(MyFlakyTest.java:22)
```
| 1.0 | Flaky test: playground.stianst.github.io.MyFlakyTest#flaky2 - ## playground.stianst.github.io.MyFlakyTest#flaky2
[Run (pull_request)](https://github.com/stianst/playground/actions/runs/3748377827) / [Pull Request #60](https://github.com/stianst/playground/pull/60)
### Errors
```
java.lang.AssertionError
at playground.stianst.github.io.MyFlakyTest.flaky2(MyFlakyTest.java:22)
```
| non_infrastructure | flaky test playground stianst github io myflakytest playground stianst github io myflakytest errors java lang assertionerror at playground stianst github io myflakytest myflakytest java | 0 |
32,877 | 27,061,245,320 | IssuesEvent | 2023-02-13 19:58:18 | cds-snc/github-secret-scanning | https://api.github.com/repos/cds-snc/github-secret-scanning | opened | Terraform: add `alarms` module | Infrastructure | ## Description
Create an `alarms` module that contains CloudWatch alarms for:
1. Route53 health check failures.
2. Secrets alerts received from GitHub. | 1.0 | Terraform: add `alarms` module - ## Description
Create an `alarms` module that contains CloudWatch alarms for:
1. Route53 health check failures.
2. Secrets alerts received from GitHub. | infrastructure | terraform add alarms module description create an alarms module that contains cloudwatch alarms for health check failures secrets alerts received from github | 1 |
30,957 | 25,198,238,060 | IssuesEvent | 2022-11-12 20:02:00 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Microsoft.CodeAnalysis.Workspaces.Common depends on a very old version of System.Composition | Area-Infrastructure Blocked Need More Info Concept-Continuous Improvement | **Version Used**:
3.6.0-3.20207.2
**Steps to Reproduce**:
1. Create a .NETStandard2.0 project (EG: following the source generators tutorial)
2. Reference Microsoft.CodeAnalysis.Workspaces.Common package.
3. Examine System.Composition version (ether in assets file, or in the references, etc).
**Expected Behavior**:
Depends on latest System.Composition consistent with other dotnet/runtime packages. 1.4.1
**Actual Behavior**:
Depends on very old System.Composition which brings in old 1.x era packages. Which is inconsistent with other dependencies from same repo.
Bug here: https://github.com/dotnet/roslyn/blob/0a4e696986ed260c38571f81ff59c325275c374e/eng/Versions.props#L73
/cc @kevinwkt | 1.0 | Microsoft.CodeAnalysis.Workspaces.Common depends on a very old version of System.Composition - **Version Used**:
3.6.0-3.20207.2
**Steps to Reproduce**:
1. Create a .NETStandard2.0 project (EG: following the source generators tutorial)
2. Reference Microsoft.CodeAnalysis.Workspaces.Common package.
3. Examine System.Composition version (ether in assets file, or in the references, etc).
**Expected Behavior**:
Depends on latest System.Composition consistent with other dotnet/runtime packages. 1.4.1
**Actual Behavior**:
Depends on very old System.Composition which brings in old 1.x era packages. Which is inconsistent with other dependencies from same repo.
Bug here: https://github.com/dotnet/roslyn/blob/0a4e696986ed260c38571f81ff59c325275c374e/eng/Versions.props#L73
/cc @kevinwkt | infrastructure | microsoft codeanalysis workspaces common depends on a very old version of system composition version used steps to reproduce create a project eg following the source generators tutorial reference microsoft codeanalysis workspaces common package examine system composition version ether in assets file or in the references etc expected behavior depends on latest system composition consistent with other dotnet runtime packages actual behavior depends on very old system composition which brings in old x era packages which is inconsistent with other dependencies from same repo bug here cc kevinwkt | 1 |
21,377 | 14,542,241,843 | IssuesEvent | 2020-12-15 15:29:19 | robotology/QA | https://api.github.com/repos/robotology/QA | closed | Restoring Unstable Cluster Setup on iCubParis002 | infrastructure software | Hello guys,
After solving many issues we had with the hardware of our two iCubs, I was trying today to setup the red ball demo, however I found the blades status quite unstable. I might need a bit of assistance on different issues that I will try to keep in this single issue. Let me start with the following problem:
When I do `ssh icub-b1` this is what I get:
``` shell
ssh icub-b1
Linux icub-netboot 3.2.0-4-amd64 #1 SMP Debian 3.2.51-1 x86_64
The programs included with the Debian GNU/Linux system are free software;
the exact distribution terms for each program are described in the
individual files in /usr/share/doc/*/copyright.
Debian GNU/Linux comes with ABSOLUTELY NO WARRANTY, to the extent
permitted by applicable law.
Exporting iCub variables for: iCubParis02
```
After which the terminal shows:
`icub@no` instead of the expected `icub@icub-b1`
If I try to do `sudo hostname icub-b1` this is what I get:
``` shell
icub@no:~$ sudo hostname icub-b1
sudo: unable to resolve host icub-b1
```
I have the same problem with all three blades. Can you help me diagnose this and the other issues I will mention after this one is solved?
As an additional information:
When I first `ssh` on any of the blades, this is what I get (notice the `icub-netboot`)
``` bash
icub@icub-desktop:~$ ssh icub-b1
Linux icub-netboot 3.2.0-4-amd64 #1 SMP Debian 3.2.51-1 x86_64
The programs included with the Debian GNU/Linux system are free software;
the exact distribution terms for each program are described in the
individual files in /usr/share/doc/*/copyright.
Debian GNU/Linux comes with ABSOLUTELY NO WARRANTY, to the extent
permitted by applicable law.
-bash: warning: setlocale: LC_ALL: cannot change locale (fr_FR.UTF-8)
-bash: warning: setlocale: LC_ALL: cannot change locale (fr_FR.UTF-8)
Exporting iCub variables for: iCubParis02
-bash: warning: setlocale: LC_ALL: cannot change locale (fr_FR.UTF-8)
```
Which reveals also another issue we have with the `locale` settings, although I'm not sure it's related to the previous issue. One at a time I guess :)
| 1.0 | Restoring Unstable Cluster Setup on iCubParis002 - Hello guys,
After solving many issues we had with the hardware of our two iCubs, I was trying today to setup the red ball demo, however I found the blades status quite unstable. I might need a bit of assistance on different issues that I will try to keep in this single issue. Let me start with the following problem:
When I do `ssh icub-b1` this is what I get:
``` shell
ssh icub-b1
Linux icub-netboot 3.2.0-4-amd64 #1 SMP Debian 3.2.51-1 x86_64
The programs included with the Debian GNU/Linux system are free software;
the exact distribution terms for each program are described in the
individual files in /usr/share/doc/*/copyright.
Debian GNU/Linux comes with ABSOLUTELY NO WARRANTY, to the extent
permitted by applicable law.
Exporting iCub variables for: iCubParis02
```
After which the terminal shows:
`icub@no` instead of the expected `icub@icub-b1`
If I try to do `sudo hostname icub-b1` this is what I get:
``` shell
icub@no:~$ sudo hostname icub-b1
sudo: unable to resolve host icub-b1
```
I have the same problem with all three blades. Can you help me diagnose this and the other issues I will mention after this one is solved?
As an additional information:
When I first `ssh` on any of the blades, this is what I get (notice the `icub-netboot`)
``` bash
icub@icub-desktop:~$ ssh icub-b1
Linux icub-netboot 3.2.0-4-amd64 #1 SMP Debian 3.2.51-1 x86_64
The programs included with the Debian GNU/Linux system are free software;
the exact distribution terms for each program are described in the
individual files in /usr/share/doc/*/copyright.
Debian GNU/Linux comes with ABSOLUTELY NO WARRANTY, to the extent
permitted by applicable law.
-bash: warning: setlocale: LC_ALL: cannot change locale (fr_FR.UTF-8)
-bash: warning: setlocale: LC_ALL: cannot change locale (fr_FR.UTF-8)
Exporting iCub variables for: iCubParis02
-bash: warning: setlocale: LC_ALL: cannot change locale (fr_FR.UTF-8)
```
Which reveals also another issue we have with the `locale` settings, although I'm not sure it's related to the previous issue. One at a time I guess :)
| infrastructure | restoring unstable cluster setup on hello guys after solving many issues we had with the hardware of our two icubs i was trying today to setup the red ball demo however i found the blades status quite unstable i might need a bit of assistance on different issues that i will try to keep in this single issue let me start with the following problem when i do ssh icub this is what i get shell ssh icub linux icub netboot smp debian the programs included with the debian gnu linux system are free software the exact distribution terms for each program are described in the individual files in usr share doc copyright debian gnu linux comes with absolutely no warranty to the extent permitted by applicable law exporting icub variables for after which the terminal shows icub no instead of the expected icub icub if i try to do sudo hostname icub this is what i get shell icub no sudo hostname icub sudo unable to resolve host icub i have the same problem with all three blades can you help me diagnose this and the other issues i will mention after this one is solved as an additional information when i first ssh on any of the blades this is what i get notice the icub netboot bash icub icub desktop ssh icub linux icub netboot smp debian the programs included with the debian gnu linux system are free software the exact distribution terms for each program are described in the individual files in usr share doc copyright debian gnu linux comes with absolutely no warranty to the extent permitted by applicable law bash warning setlocale lc all cannot change locale fr fr utf bash warning setlocale lc all cannot change locale fr fr utf exporting icub variables for bash warning setlocale lc all cannot change locale fr fr utf which reveals also another issue we have with the locale settings although i m not sure it s related to the previous issue one at a time i guess | 1 |
321,676 | 27,546,293,667 | IssuesEvent | 2023-03-07 12:02:43 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | [Flaky Test] should not scroll page | [Type] Flaky Test | <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
should not scroll page
## Test path
`specs/editor/various/navigable-toolbar.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-03-07T12:02:42.573Z"><code>[2023-03-07T12:02:42.573Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4353563197"><code>try/dedicated-block-controls-filter</code></a>.
</summary>
```
● block toolbar (contextual: false) › should not scroll page
expect(jest.fn()).not.toHaveErrored(expected)
Expected mock function not to be called but it was called with:
["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."]
at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:32:4)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| 1.0 | [Flaky Test] should not scroll page - <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
should not scroll page
## Test path
`specs/editor/various/navigable-toolbar.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-03-07T12:02:42.573Z"><code>[2023-03-07T12:02:42.573Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4353563197"><code>try/dedicated-block-controls-filter</code></a>.
</summary>
```
● block toolbar (contextual: false) › should not scroll page
expect(jest.fn()).not.toHaveErrored(expected)
Expected mock function not to be called but it was called with:
["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."]
at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:32:4)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| non_infrastructure | should not scroll page flaky test detected this is an auto generated issue by github actions please do not edit this manually test title should not scroll page test path specs editor various navigable toolbar test js errors test passed after failed attempt on a href ● block toolbar contextual false › should not scroll page expect jest fn not tohaveerrored expected expected mock function not to be called but it was called with at object assertexpectedcalls jest console build wordpress jest console src index js | 0 |
6,657 | 6,547,984,439 | IssuesEvent | 2017-09-04 17:55:39 | SatelliteQE/robottelo | https://api.github.com/repos/SatelliteQE/robottelo | reopened | Many UI tests fails to save screenshot when using saucelabs | 6.3 Infrastructure test-failure | many UI tests failed in tearDown when saving screenshot with the same error:
```python
robottelo/test.py:622: in take_screenshot
self.browser.save_screenshot(path)
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:758: in get_screenshot_as_file
png = self.get_screenshot_as_png()
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:777: in get_screenshot_as_png
return base64.b64decode(self.get_screenshot_as_base64().encode('ascii'))
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:787: in get_screenshot_as_base64
return self.execute(Command.SCREENSHOT)['value']
robottelo/ui/browser.py:44: in execute
params)
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:201: in execute
self.error_handler.check_response(response)
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/errorhandler.py:102: in check_response
value = json.loads(value_json)
/usr/lib64/python2.7/json/__init__.py:339: in loads
return _default_decoder.decode(s)
/usr/lib64/python2.7/json/decoder.py:364: in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <json.decoder.JSONDecoder object at 0x7f71580bb590>
s = "The test with session id b08625539bb64e46af50a676d3ad4345 has already finished, and can't receive further commands.
Y.../b08625539bb64e46af50a676d3ad4345
For help, please check https://wiki.saucelabs.com/display/DOCS/Common+Error+Messages"
idx = 0
def raw_decode(self, s, idx=0):
"""Decode a JSON document from ``s`` (a ``str`` or ``unicode``
beginning with a JSON document) and return a 2-tuple of the Python
representation and the index in ``s`` where the document ended.
This can be used to decode a JSON document from a string that may
have extraneous data at the end.
"""
try:
obj, end = self.scan_once(s, idx)
except StopIteration:
> raise ValueError("No JSON object could be decoded")
E ValueError: No JSON object could be decoded
/usr/lib64/python2.7/json/decoder.py:382: ValueError
```
| 1.0 | Many UI tests fails to save screenshot when using saucelabs - many UI tests failed in tearDown when saving screenshot with the same error:
```python
robottelo/test.py:622: in take_screenshot
self.browser.save_screenshot(path)
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:758: in get_screenshot_as_file
png = self.get_screenshot_as_png()
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:777: in get_screenshot_as_png
return base64.b64decode(self.get_screenshot_as_base64().encode('ascii'))
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:787: in get_screenshot_as_base64
return self.execute(Command.SCREENSHOT)['value']
robottelo/ui/browser.py:44: in execute
params)
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/webdriver.py:201: in execute
self.error_handler.check_response(response)
../../shiningpanda/jobs/375dbdea/virtualenvs/d41d8cd9/lib/python2.7/site-packages/selenium/webdriver/remote/errorhandler.py:102: in check_response
value = json.loads(value_json)
/usr/lib64/python2.7/json/__init__.py:339: in loads
return _default_decoder.decode(s)
/usr/lib64/python2.7/json/decoder.py:364: in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <json.decoder.JSONDecoder object at 0x7f71580bb590>
s = "The test with session id b08625539bb64e46af50a676d3ad4345 has already finished, and can't receive further commands.
Y.../b08625539bb64e46af50a676d3ad4345
For help, please check https://wiki.saucelabs.com/display/DOCS/Common+Error+Messages"
idx = 0
def raw_decode(self, s, idx=0):
"""Decode a JSON document from ``s`` (a ``str`` or ``unicode``
beginning with a JSON document) and return a 2-tuple of the Python
representation and the index in ``s`` where the document ended.
This can be used to decode a JSON document from a string that may
have extraneous data at the end.
"""
try:
obj, end = self.scan_once(s, idx)
except StopIteration:
> raise ValueError("No JSON object could be decoded")
E ValueError: No JSON object could be decoded
/usr/lib64/python2.7/json/decoder.py:382: ValueError
```
| infrastructure | many ui tests fails to save screenshot when using saucelabs many ui tests failed in teardown when saving screenshot with the same error python robottelo test py in take screenshot self browser save screenshot path shiningpanda jobs virtualenvs lib site packages selenium webdriver remote webdriver py in get screenshot as file png self get screenshot as png shiningpanda jobs virtualenvs lib site packages selenium webdriver remote webdriver py in get screenshot as png return self get screenshot as encode ascii shiningpanda jobs virtualenvs lib site packages selenium webdriver remote webdriver py in get screenshot as return self execute command screenshot robottelo ui browser py in execute params shiningpanda jobs virtualenvs lib site packages selenium webdriver remote webdriver py in execute self error handler check response response shiningpanda jobs virtualenvs lib site packages selenium webdriver remote errorhandler py in check response value json loads value json usr json init py in loads return default decoder decode s usr json decoder py in decode obj end self raw decode s idx w s end self s the test with session id has already finished and can t receive further commands y for help please check idx def raw decode self s idx decode a json document from s a str or unicode beginning with a json document and return a tuple of the python representation and the index in s where the document ended this can be used to decode a json document from a string that may have extraneous data at the end try obj end self scan once s idx except stopiteration raise valueerror no json object could be decoded e valueerror no json object could be decoded usr json decoder py valueerror | 1 |
27,997 | 22,711,189,567 | IssuesEvent | 2022-07-05 19:33:40 | JetBrains/compose-jb | https://api.github.com/repos/JetBrains/compose-jb | reopened | Material icons extended only being published for desktop | multiplatform infrastructure | I was having trouble using the extended material icons in my common code. It turns out it is only being built/published for desktop.
You can verify here:
https://maven.pkg.jetbrains.space/public/p/compose/dev/org/jetbrains/compose/material/
Material icons core does not have this issue.
| 1.0 | Material icons extended only being published for desktop - I was having trouble using the extended material icons in my common code. It turns out it is only being built/published for desktop.
You can verify here:
https://maven.pkg.jetbrains.space/public/p/compose/dev/org/jetbrains/compose/material/
Material icons core does not have this issue.
| infrastructure | material icons extended only being published for desktop i was having trouble using the extended material icons in my common code it turns out it is only being built published for desktop you can verify here material icons core does not have this issue | 1 |
33,688 | 27,718,203,728 | IssuesEvent | 2023-03-14 18:24:19 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | There was no runtime pack for Microsoft.AspNetCore.App available for the specified RuntimeIdentifier 'browser-wasm'. | area-infrastructure | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
@javiercn I believe that closing #45031 is an inappropriate close.
1) The three project references that are mentioned are all in another WASM project that is successfully built/run
2) The reference to obj\project.assets.json is not particularly useful. It is a a 2600 line file. If I knew what to search for it might be useful but the reference to the failing project ICEBG.Blazor shows nothing extraordinary.
3) I think you should consider conferring with @mkArtakMSFT who opened #28478 for this very issue which went on to be resolved. If it was resolved then the behavior that I am seeing is a regression. I can see that @pranavkm did the original work.
4) Since the issue is with dotnet tooling I would think you could do the research as to where the issue needs to get moved. You can't possibly expect someone outside of Microsoft to know how things are structured.
5) And it is absolutely the responsibility of the tooling to provide adequate error messages. Right now it is showing the equivalent of 'build broken -- good luck"
Needless to say I am very disappointed in how this issue has been handled.
### Expected Behavior
See #45031
### Steps To Reproduce
See #45031
### Exceptions (if any)
_No response_
### .NET Version
_No response_
### Anything else?
_No response_ | 1.0 | There was no runtime pack for Microsoft.AspNetCore.App available for the specified RuntimeIdentifier 'browser-wasm'. - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
@javiercn I believe that closing #45031 is an inappropriate close.
1) The three project references that are mentioned are all in another WASM project that is successfully built/run
2) The reference to obj\project.assets.json is not particularly useful. It is a a 2600 line file. If I knew what to search for it might be useful but the reference to the failing project ICEBG.Blazor shows nothing extraordinary.
3) I think you should consider conferring with @mkArtakMSFT who opened #28478 for this very issue which went on to be resolved. If it was resolved then the behavior that I am seeing is a regression. I can see that @pranavkm did the original work.
4) Since the issue is with dotnet tooling I would think you could do the research as to where the issue needs to get moved. You can't possibly expect someone outside of Microsoft to know how things are structured.
5) And it is absolutely the responsibility of the tooling to provide adequate error messages. Right now it is showing the equivalent of 'build broken -- good luck"
Needless to say I am very disappointed in how this issue has been handled.
### Expected Behavior
See #45031
### Steps To Reproduce
See #45031
### Exceptions (if any)
_No response_
### .NET Version
_No response_
### Anything else?
_No response_ | infrastructure | there was no runtime pack for microsoft aspnetcore app available for the specified runtimeidentifier browser wasm is there an existing issue for this i have searched the existing issues describe the bug javiercn i believe that closing is an inappropriate close the three project references that are mentioned are all in another wasm project that is successfully built run the reference to obj project assets json is not particularly useful it is a a line file if i knew what to search for it might be useful but the reference to the failing project icebg blazor shows nothing extraordinary i think you should consider conferring with mkartakmsft who opened for this very issue which went on to be resolved if it was resolved then the behavior that i am seeing is a regression i can see that pranavkm did the original work since the issue is with dotnet tooling i would think you could do the research as to where the issue needs to get moved you can t possibly expect someone outside of microsoft to know how things are structured and it is absolutely the responsibility of the tooling to provide adequate error messages right now it is showing the equivalent of build broken good luck needless to say i am very disappointed in how this issue has been handled expected behavior see steps to reproduce see exceptions if any no response net version no response anything else no response | 1 |
13,825 | 5,468,080,436 | IssuesEvent | 2017-03-10 04:06:58 | docker/docker | https://api.github.com/repos/docker/docker | closed | Exporting volume from container running in windows 10 insider preview doesn't work | area/builder area/bundles platform/windows version/1.13 | **Description**
I am trying the latest windows 10 insider preview build with docker overlay driver support.
I have a image thats build with the below `Dockerfile`
```
FROM microsoft/windowsservercore
COPY . C:/temp
SHELL ["powershell", "-Command"]
RUN New-Item c:\Qumram -Type directory; \
New-Item C:\Qumram\start.ps1 -Type File -Value \
'Param( \
[string]$LogFilePath = \"C:\Qumram\config\config.log\" \
) \
Copy-Item C:\temp\* C:\Qumram\config -Recurse | Out-Null; \
New-Item -Path $LogFilePath -Itemtype file -Force | Out-Null; \
Get-Content -Path $LogFilePath -Wait;'
VOLUME C:\\Qumram\\config
SHELL ["powershell", "-File"]
HEALTHCHECK CMD exit 0
ENTRYPOINT C:\Qumram\start.ps1
```
And i deploy a stack using the below `compose.yml` file
```
version: '3.0'
volumes:
windows-configuration:
services:
config-windows-volume:
image: josejiby/qumram:testconfig-windows-dev
volumes:
- windows-configuration:\\c\Qumram\config
deploy:
mode: replicated
replicas: 1
placement:
constraints:
- engine.labels.type == windows
restart_policy:
condition: on-failure
```
I used to use `volumes_from` in the earlier 2.0 format in windowsservercore to share the volume, however i am using the new v3 format hence the named volume in compose file.
**Steps to reproduce the issue:**
1. I run `docker -H linuxhost stack deploy -c compose.yml test
2. It schedules and starts running the container
3. The container fails throwing an error `starting container failed: container a85162e691c88570cc097eb709862659838d6e68291bf55f6aa1672b887bbc7c encountered an error during Start: failure in a Windows system call: The compute system exited unexpectedly. (0xc0370106)`
**Describe the results you received:**
The container should have started exporting the folder as a named volume
**Describe the results you expected:**
The container crashes
**Additional information you deem important (e.g. issue happens only occasionally):**
**Output of `docker version`:**
```
Client:
Version: 1.13.1
API version: 1.26
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: darwin/amd64
Server:
Version: 1.13.1
API version: 1.26 (minimum version 1.24)
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: windows/amd64
Experimental: false
```
**Output of `docker info`:**
```
Containers: 5
Running: 0
Paused: 0
Stopped: 5
Images: 57
Server Version: 1.13.1
Storage Driver: windowsfilter
Windows:
Logging Driver: json-file
Plugins:
Volume: local
Network: l2bridge l2tunnel nat null overlay transparent
Swarm: active
NodeID: mzpe8u9g1448u9aw7ph2s8wub
Is Manager: false
Node Address: 10.211.55.7
Manager Addresses:
10.211.55.8:2377
Default Isolation: hyperv
Kernel Version: 10.0 15031 (15031.0.amd64fre.rs2_release.170204-1546)
Operating System: Windows 10 Enterprise Insider Preview
OSType: windows
Architecture: x86_64
CPUs: 1
Total Memory: 1.983 GiB
Name: windows
ID: 5262:FAUI:VLWB:COL4:Z3HH:5XGE:DBPD:HFBS:VYXB:CNAY:VKV3:CGFT
Docker Root Dir: C:\ProgramData\docker
Debug Mode (client): false
Debug Mode (server): false
Username: josejiby
Registry: https://index.docker.io/v1/
Labels:
type=windows
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
The windows host is running inside parallels along side an ubuntu vm, which is the master of the swarm.
| 1.0 | Exporting volume from container running in windows 10 insider preview doesn't work - **Description**
I am trying the latest windows 10 insider preview build with docker overlay driver support.
I have a image thats build with the below `Dockerfile`
```
FROM microsoft/windowsservercore
COPY . C:/temp
SHELL ["powershell", "-Command"]
RUN New-Item c:\Qumram -Type directory; \
New-Item C:\Qumram\start.ps1 -Type File -Value \
'Param( \
[string]$LogFilePath = \"C:\Qumram\config\config.log\" \
) \
Copy-Item C:\temp\* C:\Qumram\config -Recurse | Out-Null; \
New-Item -Path $LogFilePath -Itemtype file -Force | Out-Null; \
Get-Content -Path $LogFilePath -Wait;'
VOLUME C:\\Qumram\\config
SHELL ["powershell", "-File"]
HEALTHCHECK CMD exit 0
ENTRYPOINT C:\Qumram\start.ps1
```
And i deploy a stack using the below `compose.yml` file
```
version: '3.0'
volumes:
windows-configuration:
services:
config-windows-volume:
image: josejiby/qumram:testconfig-windows-dev
volumes:
- windows-configuration:\\c\Qumram\config
deploy:
mode: replicated
replicas: 1
placement:
constraints:
- engine.labels.type == windows
restart_policy:
condition: on-failure
```
I used to use `volumes_from` in the earlier 2.0 format in windowsservercore to share the volume, however i am using the new v3 format hence the named volume in compose file.
**Steps to reproduce the issue:**
1. I run `docker -H linuxhost stack deploy -c compose.yml test
2. It schedules and starts running the container
3. The container fails throwing an error `starting container failed: container a85162e691c88570cc097eb709862659838d6e68291bf55f6aa1672b887bbc7c encountered an error during Start: failure in a Windows system call: The compute system exited unexpectedly. (0xc0370106)`
**Describe the results you received:**
The container should have started exporting the folder as a named volume
**Describe the results you expected:**
The container crashes
**Additional information you deem important (e.g. issue happens only occasionally):**
**Output of `docker version`:**
```
Client:
Version: 1.13.1
API version: 1.26
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: darwin/amd64
Server:
Version: 1.13.1
API version: 1.26 (minimum version 1.24)
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: windows/amd64
Experimental: false
```
**Output of `docker info`:**
```
Containers: 5
Running: 0
Paused: 0
Stopped: 5
Images: 57
Server Version: 1.13.1
Storage Driver: windowsfilter
Windows:
Logging Driver: json-file
Plugins:
Volume: local
Network: l2bridge l2tunnel nat null overlay transparent
Swarm: active
NodeID: mzpe8u9g1448u9aw7ph2s8wub
Is Manager: false
Node Address: 10.211.55.7
Manager Addresses:
10.211.55.8:2377
Default Isolation: hyperv
Kernel Version: 10.0 15031 (15031.0.amd64fre.rs2_release.170204-1546)
Operating System: Windows 10 Enterprise Insider Preview
OSType: windows
Architecture: x86_64
CPUs: 1
Total Memory: 1.983 GiB
Name: windows
ID: 5262:FAUI:VLWB:COL4:Z3HH:5XGE:DBPD:HFBS:VYXB:CNAY:VKV3:CGFT
Docker Root Dir: C:\ProgramData\docker
Debug Mode (client): false
Debug Mode (server): false
Username: josejiby
Registry: https://index.docker.io/v1/
Labels:
type=windows
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
The windows host is running inside parallels along side an ubuntu vm, which is the master of the swarm.
| non_infrastructure | exporting volume from container running in windows insider preview doesn t work description i am trying the latest windows insider preview build with docker overlay driver support i have a image thats build with the below dockerfile from microsoft windowsservercore copy c temp shell run new item c qumram type directory new item c qumram start type file value param logfilepath c qumram config config log copy item c temp c qumram config recurse out null new item path logfilepath itemtype file force out null get content path logfilepath wait volume c qumram config shell healthcheck cmd exit entrypoint c qumram start and i deploy a stack using the below compose yml file version volumes windows configuration services config windows volume image josejiby qumram testconfig windows dev volumes windows configuration c qumram config deploy mode replicated replicas placement constraints engine labels type windows restart policy condition on failure i used to use volumes from in the earlier format in windowsservercore to share the volume however i am using the new format hence the named volume in compose file steps to reproduce the issue i run docker h linuxhost stack deploy c compose yml test it schedules and starts running the container the container fails throwing an error starting container failed container encountered an error during start failure in a windows system call the compute system exited unexpectedly describe the results you received the container should have started exporting the folder as a named volume describe the results you expected the container crashes additional information you deem important e g issue happens only occasionally output of docker version client version api version go version git commit built wed feb os arch darwin server version api version minimum version go version git commit built wed feb os arch windows experimental false output of docker info containers running paused stopped images server version storage driver windowsfilter windows logging driver json file plugins volume local network nat null overlay transparent swarm active nodeid is manager false node address manager addresses default isolation hyperv kernel version release operating system windows enterprise insider preview ostype windows architecture cpus total memory gib name windows id faui vlwb dbpd hfbs vyxb cnay cgft docker root dir c programdata docker debug mode client false debug mode server false username josejiby registry labels type windows experimental false insecure registries live restore enabled false additional environment details aws virtualbox physical etc the windows host is running inside parallels along side an ubuntu vm which is the master of the swarm | 0 |
591,221 | 17,797,861,453 | IssuesEvent | 2021-09-01 01:56:13 | LuPro/SlabelFish | https://api.github.com/repos/LuPro/SlabelFish | closed | It's not JSON output | high priority | The output is definitely human-readable, and while it looks _like_ JSON it is unfortunately not valid JSON.
For example python uses single quotes when printing strings and JSON requires double qoutes.
## Example
```
>>> import json
>>> me = {"name": "flofriday", "age": 42}
>>> print(me)
{'name': 'flofriday', 'age': 42}
>>> print(json.dumps(me))
{"name": "flofriday", "age": 42}
```
## Fix
Use `print(json.dumps(assets))` | 1.0 | It's not JSON output - The output is definitely human-readable, and while it looks _like_ JSON it is unfortunately not valid JSON.
For example python uses single quotes when printing strings and JSON requires double qoutes.
## Example
```
>>> import json
>>> me = {"name": "flofriday", "age": 42}
>>> print(me)
{'name': 'flofriday', 'age': 42}
>>> print(json.dumps(me))
{"name": "flofriday", "age": 42}
```
## Fix
Use `print(json.dumps(assets))` | non_infrastructure | it s not json output the output is definitely human readable and while it looks like json it is unfortunately not valid json for example python uses single quotes when printing strings and json requires double qoutes example import json me name flofriday age print me name flofriday age print json dumps me name flofriday age fix use print json dumps assets | 0 |
2,650 | 8,102,838,058 | IssuesEvent | 2018-08-13 04:48:56 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | Jenkins is becoming Idle for pipeline build in OSIO launcher flow. | SEV2-high area/architecture/build priority/P4 sprint/next team/build-cd type/bug | Due to this Jenkins issue, No build could not able to see the finish line.
This is a critical issue from the build pipeline endpoint. Please check the below screenshot.

| 1.0 | Jenkins is becoming Idle for pipeline build in OSIO launcher flow. - Due to this Jenkins issue, No build could not able to see the finish line.
This is a critical issue from the build pipeline endpoint. Please check the below screenshot.

| non_infrastructure | jenkins is becoming idle for pipeline build in osio launcher flow due to this jenkins issue no build could not able to see the finish line this is a critical issue from the build pipeline endpoint please check the below screenshot | 0 |
223,237 | 17,575,347,874 | IssuesEvent | 2021-08-15 13:55:58 | summercms/sc-user-interface-module | https://api.github.com/repos/summercms/sc-user-interface-module | closed | Setup ui settings section and add icon package selector | enhancement Dashboard FINSIHED Priority: Medium Testing - Passed | ### Enhancement idea
- [x] Setup ui settings section and add icon selector.
| 1.0 | Setup ui settings section and add icon package selector - ### Enhancement idea
- [x] Setup ui settings section and add icon selector.
| non_infrastructure | setup ui settings section and add icon package selector enhancement idea setup ui settings section and add icon selector | 0 |
413,705 | 12,090,754,686 | IssuesEvent | 2020-04-19 08:17:43 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | [MSVC Static] DeviceGuardImpl for CPU is not available | high priority module: internals module: windows topic: static linking triaged | ## 🐛 Bug
After having built a static libtorch.a and loading a model, I get the following exception:
c10::Error: DeviceGuardImpl for CPU is not available.
## To Reproduce
Steps to reproduce the behavior:
1. Build static libtorch on Windows
2. Attempt to load model
`std::shared_ptr<torch::jit::script::Module> model = torch::jit::load(model_path, torch::kCUDA);`
## Expected behavior
Works
## Environment
- PyTorch Version (e.g., 1.0): master
- OS (e.g., Linux): Windows
- How you installed PyTorch (`conda`, `pip`, source): source
- Build command you used (if compiling from source): cmake -DBUILD_TORCH=ON -DTORCH_STATIC=ON -DBUILD_SHARED_LIBS=OFF
- Python version: 3.7
- CUDA/cuDNN version: 10.1/7.5
- GPU models and configuration: 1060
- Any other relevant information: | 1.0 | [MSVC Static] DeviceGuardImpl for CPU is not available - ## 🐛 Bug
After having built a static libtorch.a and loading a model, I get the following exception:
c10::Error: DeviceGuardImpl for CPU is not available.
## To Reproduce
Steps to reproduce the behavior:
1. Build static libtorch on Windows
2. Attempt to load model
`std::shared_ptr<torch::jit::script::Module> model = torch::jit::load(model_path, torch::kCUDA);`
## Expected behavior
Works
## Environment
- PyTorch Version (e.g., 1.0): master
- OS (e.g., Linux): Windows
- How you installed PyTorch (`conda`, `pip`, source): source
- Build command you used (if compiling from source): cmake -DBUILD_TORCH=ON -DTORCH_STATIC=ON -DBUILD_SHARED_LIBS=OFF
- Python version: 3.7
- CUDA/cuDNN version: 10.1/7.5
- GPU models and configuration: 1060
- Any other relevant information: | non_infrastructure | deviceguardimpl for cpu is not available 🐛 bug after having built a static libtorch a and loading a model i get the following exception error deviceguardimpl for cpu is not available to reproduce steps to reproduce the behavior build static libtorch on windows attempt to load model std shared ptr model torch jit load model path torch kcuda expected behavior works environment pytorch version e g master os e g linux windows how you installed pytorch conda pip source source build command you used if compiling from source cmake dbuild torch on dtorch static on dbuild shared libs off python version cuda cudnn version gpu models and configuration any other relevant information | 0 |
42,358 | 17,106,124,130 | IssuesEvent | 2021-07-09 18:02:33 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | Windows FSx DNS alias support | enhancement service/fsx | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
AWS recently added support for DNS aliases for Windows FSx systems. Please add this as an available attribute in in the aws_fsx_windows_file_system resource.
https://aws.amazon.com/about-aws/whats-new/2020/11/amazon-fsx-windows-file-server-supports-access-file-systems-alternate-dns-names/
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* aws_fsx_windows_file_system
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "aws_fsx_windows_file_system" "example" {
active_directory_id = aws_directory_service_directory.example.id
kms_key_id = aws_kms_key.example.arn
storage_capacity = 300
subnet_ids = [aws_subnet.example.id]
throughput_capacity = 1024
dns_alias = ["fsx.domain.local"]
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
* #0000
| 1.0 | Windows FSx DNS alias support - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
AWS recently added support for DNS aliases for Windows FSx systems. Please add this as an available attribute in in the aws_fsx_windows_file_system resource.
https://aws.amazon.com/about-aws/whats-new/2020/11/amazon-fsx-windows-file-server-supports-access-file-systems-alternate-dns-names/
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* aws_fsx_windows_file_system
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "aws_fsx_windows_file_system" "example" {
active_directory_id = aws_directory_service_directory.example.id
kms_key_id = aws_kms_key.example.arn
storage_capacity = 300
subnet_ids = [aws_subnet.example.id]
throughput_capacity = 1024
dns_alias = ["fsx.domain.local"]
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
* #0000
| non_infrastructure | windows fsx dns alias support community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description aws recently added support for dns aliases for windows fsx systems please add this as an available attribute in in the aws fsx windows file system resource new or affected resource s aws fsx windows file system potential terraform configuration hcl resource aws fsx windows file system example active directory id aws directory service directory example id kms key id aws kms key example arn storage capacity subnet ids throughput capacity dns alias references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example | 0 |
298,163 | 9,196,440,717 | IssuesEvent | 2019-03-07 07:04:59 | servicemesher/istio-official-translation | https://api.github.com/repos/servicemesher/istio-official-translation | closed | /docs/tasks/telemetry/fluentd/index.md | finished lang/zh priority/P0 sync/update version/1.1 | Source File: [/docs/tasks/telemetry/fluentd/index.md](https://github.com/istio/istio.io/tree/master/content/docs/tasks/telemetry/fluentd/index.md)
Diff:
~~~diff
diff --git a/content/docs/tasks/telemetry/fluentd/index.md b/content/docs/tasks/telemetry/fluentd/index.md
index 6059c6f7..9569b94b 100644
--- a/content/docs/tasks/telemetry/fluentd/index.md
+++ b/content/docs/tasks/telemetry/fluentd/index.md
@@ -315,7 +315,7 @@ spec:
latency: response.duration | "0ms"
monitored_resource_type: '"UNSPECIFIED"'
---
-# Configuration for a Fluentd handler
+# Configuration for a fluentd handler
apiVersion: "config.istio.io/v1alpha2"
kind: fluentd
metadata:
@@ -324,7 +324,7 @@ metadata:
spec:
address: "fluentd-es.logging:24224"
---
-# Rule to send logentry instances to the Fluentd handler
+# Rule to send logentry instances to the fluentd handler
apiVersion: "config.istio.io/v1alpha2"
kind: rule
metadata:
~~~ | 1.0 | /docs/tasks/telemetry/fluentd/index.md - Source File: [/docs/tasks/telemetry/fluentd/index.md](https://github.com/istio/istio.io/tree/master/content/docs/tasks/telemetry/fluentd/index.md)
Diff:
~~~diff
diff --git a/content/docs/tasks/telemetry/fluentd/index.md b/content/docs/tasks/telemetry/fluentd/index.md
index 6059c6f7..9569b94b 100644
--- a/content/docs/tasks/telemetry/fluentd/index.md
+++ b/content/docs/tasks/telemetry/fluentd/index.md
@@ -315,7 +315,7 @@ spec:
latency: response.duration | "0ms"
monitored_resource_type: '"UNSPECIFIED"'
---
-# Configuration for a Fluentd handler
+# Configuration for a fluentd handler
apiVersion: "config.istio.io/v1alpha2"
kind: fluentd
metadata:
@@ -324,7 +324,7 @@ metadata:
spec:
address: "fluentd-es.logging:24224"
---
-# Rule to send logentry instances to the Fluentd handler
+# Rule to send logentry instances to the fluentd handler
apiVersion: "config.istio.io/v1alpha2"
kind: rule
metadata:
~~~ | non_infrastructure | docs tasks telemetry fluentd index md source file diff diff diff git a content docs tasks telemetry fluentd index md b content docs tasks telemetry fluentd index md index a content docs tasks telemetry fluentd index md b content docs tasks telemetry fluentd index md spec latency response duration monitored resource type unspecified configuration for a fluentd handler configuration for a fluentd handler apiversion config istio io kind fluentd metadata metadata spec address fluentd es logging rule to send logentry instances to the fluentd handler rule to send logentry instances to the fluentd handler apiversion config istio io kind rule metadata | 0 |
204,329 | 23,239,491,446 | IssuesEvent | 2022-08-03 14:28:58 | turkdevops/angular | https://api.github.com/repos/turkdevops/angular | closed | CVE-2020-7768 (High) detected in multiple libraries - autoclosed | security vulnerability | ## CVE-2020-7768 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>grpc-js-0.6.16.tgz</b>, <b>grpc-js-0.6.18.tgz</b>, <b>grpc-js-0.7.5.tgz</b>, <b>grpc-js-1.0.5.tgz</b></p></summary>
<p>
<details><summary><b>grpc-js-0.6.16.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.16.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.16.tgz</a></p>
<p>Path to dependency file: /aio/package.json</p>
<p>Path to vulnerable library: /aio/node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-8.14.1.tgz (Root Library)
- google-gax-1.12.0.tgz
- :x: **grpc-js-0.6.16.tgz** (Vulnerable Library)
</details>
<details><summary><b>grpc-js-0.6.18.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.18.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.18.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-7.16.2.tgz (Root Library)
- google-gax-1.12.0.tgz
- :x: **grpc-js-0.6.18.tgz** (Vulnerable Library)
</details>
<details><summary><b>grpc-js-0.7.5.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.7.5.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.7.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-7.16.2.tgz (Root Library)
- pubsub-1.7.0.tgz
- google-gax-1.15.2.tgz
- :x: **grpc-js-0.7.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>grpc-js-1.0.5.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.0.5.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.0.5.tgz</a></p>
<p>Path to dependency file: /aio/package.json</p>
<p>Path to vulnerable library: /aio/node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-8.14.1.tgz (Root Library)
- pubsub-1.7.3.tgz
- google-gax-1.15.3.tgz
- :x: **grpc-js-1.0.5.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/angular/commit/c6aca37f442da8c55a02d7c53ccc58100ab004f3">c6aca37f442da8c55a02d7c53ccc58100ab004f3</a></p>
<p>Found in base branch: <b>labs/router</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package grpc before 1.24.4; the package @grpc/grpc-js before 1.1.8 are vulnerable to Prototype Pollution via loadPackageDefinition.
<p>Publish Date: 2020-11-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7768>CVE-2020-7768</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7768">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7768</a></p>
<p>Release Date: 2020-11-11</p>
<p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p><p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p><p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p><p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7768 (High) detected in multiple libraries - autoclosed - ## CVE-2020-7768 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>grpc-js-0.6.16.tgz</b>, <b>grpc-js-0.6.18.tgz</b>, <b>grpc-js-0.7.5.tgz</b>, <b>grpc-js-1.0.5.tgz</b></p></summary>
<p>
<details><summary><b>grpc-js-0.6.16.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.16.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.16.tgz</a></p>
<p>Path to dependency file: /aio/package.json</p>
<p>Path to vulnerable library: /aio/node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-8.14.1.tgz (Root Library)
- google-gax-1.12.0.tgz
- :x: **grpc-js-0.6.16.tgz** (Vulnerable Library)
</details>
<details><summary><b>grpc-js-0.6.18.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.18.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.6.18.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-7.16.2.tgz (Root Library)
- google-gax-1.12.0.tgz
- :x: **grpc-js-0.6.18.tgz** (Vulnerable Library)
</details>
<details><summary><b>grpc-js-0.7.5.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.7.5.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-0.7.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-7.16.2.tgz (Root Library)
- pubsub-1.7.0.tgz
- google-gax-1.15.2.tgz
- :x: **grpc-js-0.7.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>grpc-js-1.0.5.tgz</b></p></summary>
<p>gRPC Library for Node - pure JS implementation</p>
<p>Library home page: <a href="https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.0.5.tgz">https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.0.5.tgz</a></p>
<p>Path to dependency file: /aio/package.json</p>
<p>Path to vulnerable library: /aio/node_modules/@grpc/grpc-js/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-8.14.1.tgz (Root Library)
- pubsub-1.7.3.tgz
- google-gax-1.15.3.tgz
- :x: **grpc-js-1.0.5.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/angular/commit/c6aca37f442da8c55a02d7c53ccc58100ab004f3">c6aca37f442da8c55a02d7c53ccc58100ab004f3</a></p>
<p>Found in base branch: <b>labs/router</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package grpc before 1.24.4; the package @grpc/grpc-js before 1.1.8 are vulnerable to Prototype Pollution via loadPackageDefinition.
<p>Publish Date: 2020-11-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7768>CVE-2020-7768</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7768">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7768</a></p>
<p>Release Date: 2020-11-11</p>
<p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p><p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p><p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p><p>Fix Resolution (@grpc/grpc-js): 1.1.8</p>
<p>Direct dependency fix Resolution (firebase-tools): 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in multiple libraries autoclosed cve high severity vulnerability vulnerable libraries grpc js tgz grpc js tgz grpc js tgz grpc js tgz grpc js tgz grpc library for node pure js implementation library home page a href path to dependency file aio package json path to vulnerable library aio node modules grpc grpc js package json dependency hierarchy firebase tools tgz root library google gax tgz x grpc js tgz vulnerable library grpc js tgz grpc library for node pure js implementation library home page a href path to dependency file package json path to vulnerable library node modules grpc grpc js package json dependency hierarchy firebase tools tgz root library google gax tgz x grpc js tgz vulnerable library grpc js tgz grpc library for node pure js implementation library home page a href path to dependency file package json path to vulnerable library node modules grpc grpc js package json dependency hierarchy firebase tools tgz root library pubsub tgz google gax tgz x grpc js tgz vulnerable library grpc js tgz grpc library for node pure js implementation library home page a href path to dependency file aio package json path to vulnerable library aio node modules grpc grpc js package json dependency hierarchy firebase tools tgz root library pubsub tgz google gax tgz x grpc js tgz vulnerable library found in head commit a href found in base branch labs router vulnerability details the package grpc before the package grpc grpc js before are vulnerable to prototype pollution via loadpackagedefinition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution grpc grpc js direct dependency fix resolution firebase tools fix resolution grpc grpc js direct dependency fix resolution firebase tools fix resolution grpc grpc js direct dependency fix resolution firebase tools fix resolution grpc grpc js direct dependency fix resolution firebase tools step up your open source security game with mend | 0 |
298,641 | 25,841,959,741 | IssuesEvent | 2022-12-13 01:32:15 | celestiaorg/celestia-app | https://api.github.com/repos/celestiaorg/celestia-app | opened | Run long fuzzing tests per PR after they are merged | github_actions testing | we should have longer running fuzzing tests that run after a PR is merged to hopefully catch bugs such as that found in #1108 | 1.0 | Run long fuzzing tests per PR after they are merged - we should have longer running fuzzing tests that run after a PR is merged to hopefully catch bugs such as that found in #1108 | non_infrastructure | run long fuzzing tests per pr after they are merged we should have longer running fuzzing tests that run after a pr is merged to hopefully catch bugs such as that found in | 0 |
55,562 | 6,486,156,937 | IssuesEvent | 2017-08-19 17:05:55 | TheProgrammingKitchen/apm | https://api.github.com/repos/TheProgrammingKitchen/apm | closed | Confirm delete and Integration test for Delete Issue | enhancement Technical Dept tests | Delete issue works without user-confirmation at the moment. Add a pop-up to ask the user to confirm the delete action.
Write an integration test for this usecase. | 1.0 | Confirm delete and Integration test for Delete Issue - Delete issue works without user-confirmation at the moment. Add a pop-up to ask the user to confirm the delete action.
Write an integration test for this usecase. | non_infrastructure | confirm delete and integration test for delete issue delete issue works without user confirmation at the moment add a pop up to ask the user to confirm the delete action write an integration test for this usecase | 0 |
9,845 | 3,074,964,853 | IssuesEvent | 2015-08-20 10:42:31 | medic/medic-webapp | https://api.github.com/repos/medic/medic-webapp | closed | Requesting futon does not show basic auth prompt | 3 - Acceptance testing Bug | The only way to log in is via the Medic login page. This is problematic because if the login page is unavailable for some reason we need to get access to fix it.
This is probably because all requests are being routed through api. We should make sure that _utils, _users, dashboard, etc are available via basic auth prompt. | 1.0 | Requesting futon does not show basic auth prompt - The only way to log in is via the Medic login page. This is problematic because if the login page is unavailable for some reason we need to get access to fix it.
This is probably because all requests are being routed through api. We should make sure that _utils, _users, dashboard, etc are available via basic auth prompt. | non_infrastructure | requesting futon does not show basic auth prompt the only way to log in is via the medic login page this is problematic because if the login page is unavailable for some reason we need to get access to fix it this is probably because all requests are being routed through api we should make sure that utils users dashboard etc are available via basic auth prompt | 0 |
25,726 | 4,167,015,289 | IssuesEvent | 2016-06-20 07:44:14 | Beit-Hatfutsot/dbs-front | https://api.github.com/repos/Beit-Hatfutsot/dbs-front | closed | Mouse cursor turns to editing-cursor when hovered over "remove from my story". | beginner bug Need Tests | When the user hovers the mouse over the "Remove from my story" option, the mouse cursor will turn into a text-edit cursor. It should not do so.
Repro:
1. Log in
2. Arrive at my story
3. Remove an item from my story.
Result:
When attempting to remove the item, the cursor shifts to editing mode. | 1.0 | Mouse cursor turns to editing-cursor when hovered over "remove from my story". - When the user hovers the mouse over the "Remove from my story" option, the mouse cursor will turn into a text-edit cursor. It should not do so.
Repro:
1. Log in
2. Arrive at my story
3. Remove an item from my story.
Result:
When attempting to remove the item, the cursor shifts to editing mode. | non_infrastructure | mouse cursor turns to editing cursor when hovered over remove from my story when the user hovers the mouse over the remove from my story option the mouse cursor will turn into a text edit cursor it should not do so repro log in arrive at my story remove an item from my story result when attempting to remove the item the cursor shifts to editing mode | 0 |
18,225 | 12,837,923,396 | IssuesEvent | 2020-07-07 16:33:32 | dotnet/docker-tools | https://api.github.com/repos/dotnet/docker-tools | closed | Use aarch64 version of AzDO agent | area:infrastructure enhancement triaged | The current infrastructure for supporting Linux ARM builds has been to use a Linux AMD64 agent machine that sends remote Docker commands to an aarch64 machine. This was necessary because the AzDO agent software didn't support aarch64. But now that a release candidate is available for this support (https://github.com/microsoft/azure-pipelines-agent/issues/1911), we can update the infrastructure so that the Linux AMD64 agent machine goes away and the aarch64 machine directly runs the agent software.
- [x] Publish an ARM version of [Image Builder](https://github.com/dotnet/docker-tools/blob/master/src/Microsoft.DotNet.ImageBuilder/Dockerfile.linux) with a multi-arch tag
- [x] Publish an ARM version of the [Test Runner](https://github.com/dotnet/docker-tools/blob/master/src/Microsoft.DotNet.ImageBuilder/Dockerfile.linux.runner) image with a multi-arch tag
- [x] Update common pipeline logic to consume multi-arch tags for Image Builder and Test Runner
- [x] Update common pipeline logic to conditionally support ARM agents
- [x] Install agent software on public ARM machines
- [x] Disable use of remote Docker server in common pipeline logic in nightly branch
- [x] Test for a period of time to ensure builds are working
- [ ] Install agent software on internal ARM machines
- [ ] Disable use of remote Docker server in common pipeline logic in master branch
- [ ] De-allocate AMD build agents used w/ARM machines | 1.0 | Use aarch64 version of AzDO agent - The current infrastructure for supporting Linux ARM builds has been to use a Linux AMD64 agent machine that sends remote Docker commands to an aarch64 machine. This was necessary because the AzDO agent software didn't support aarch64. But now that a release candidate is available for this support (https://github.com/microsoft/azure-pipelines-agent/issues/1911), we can update the infrastructure so that the Linux AMD64 agent machine goes away and the aarch64 machine directly runs the agent software.
- [x] Publish an ARM version of [Image Builder](https://github.com/dotnet/docker-tools/blob/master/src/Microsoft.DotNet.ImageBuilder/Dockerfile.linux) with a multi-arch tag
- [x] Publish an ARM version of the [Test Runner](https://github.com/dotnet/docker-tools/blob/master/src/Microsoft.DotNet.ImageBuilder/Dockerfile.linux.runner) image with a multi-arch tag
- [x] Update common pipeline logic to consume multi-arch tags for Image Builder and Test Runner
- [x] Update common pipeline logic to conditionally support ARM agents
- [x] Install agent software on public ARM machines
- [x] Disable use of remote Docker server in common pipeline logic in nightly branch
- [x] Test for a period of time to ensure builds are working
- [ ] Install agent software on internal ARM machines
- [ ] Disable use of remote Docker server in common pipeline logic in master branch
- [ ] De-allocate AMD build agents used w/ARM machines | infrastructure | use version of azdo agent the current infrastructure for supporting linux arm builds has been to use a linux agent machine that sends remote docker commands to an machine this was necessary because the azdo agent software didn t support but now that a release candidate is available for this support we can update the infrastructure so that the linux agent machine goes away and the machine directly runs the agent software publish an arm version of with a multi arch tag publish an arm version of the image with a multi arch tag update common pipeline logic to consume multi arch tags for image builder and test runner update common pipeline logic to conditionally support arm agents install agent software on public arm machines disable use of remote docker server in common pipeline logic in nightly branch test for a period of time to ensure builds are working install agent software on internal arm machines disable use of remote docker server in common pipeline logic in master branch de allocate amd build agents used w arm machines | 1 |
11,244 | 9,015,062,024 | IssuesEvent | 2019-02-06 00:52:36 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | Predefined type 'System.ValueTuple is not defined or imported | area-infrastructure | ASP.NET Core Web Application targeting .net 4.7.1 cannot use System.ValueTuple.
Repro:
1. Create a new ASP.NET Core Web Application that targets .net 4.7.1
2. Add a class that has a tuple
3. Observe errors.
-Cannot define a class or member that utilizes tuples because the compiler required type 'System.Runtime.CompilerServices.TupleElementNamesAttribute' cannot be found. Are you missing a reference?
-Predefined type 'System.ValueTuple`2' is not defined or imported
| 1.0 | Predefined type 'System.ValueTuple is not defined or imported - ASP.NET Core Web Application targeting .net 4.7.1 cannot use System.ValueTuple.
Repro:
1. Create a new ASP.NET Core Web Application that targets .net 4.7.1
2. Add a class that has a tuple
3. Observe errors.
-Cannot define a class or member that utilizes tuples because the compiler required type 'System.Runtime.CompilerServices.TupleElementNamesAttribute' cannot be found. Are you missing a reference?
-Predefined type 'System.ValueTuple`2' is not defined or imported
| infrastructure | predefined type system valuetuple is not defined or imported asp net core web application targeting net cannot use system valuetuple repro create a new asp net core web application that targets net add a class that has a tuple observe errors cannot define a class or member that utilizes tuples because the compiler required type system runtime compilerservices tupleelementnamesattribute cannot be found are you missing a reference predefined type system valuetuple is not defined or imported | 1 |
319,348 | 9,742,593,846 | IssuesEvent | 2019-06-02 18:25:22 | vdenotaris/spring-boot-security-saml-sample | https://api.github.com/repos/vdenotaris/spring-boot-security-saml-sample | closed | Usage of an external keystore | contrib/help-wanted kind/enhancement priority/important-soon | **Describe the bug**
Since SSOCircle changes certificates quite often, an external keystore is needed to properly separate the building stage and the running stage while using Docker images.
**To Reproduce**
Steps to reproduce the behavior:
1. The command `docker run -it --rm -p 8080:8080 -t vdenotaris/spring-saml-sp:2.1.0-openjdk8` just fails when a new certificate is deployed for SSOCircle.
**Expected behavior**
Upgrade the certificate within the running stage instead the onbuild.
| 1.0 | Usage of an external keystore - **Describe the bug**
Since SSOCircle changes certificates quite often, an external keystore is needed to properly separate the building stage and the running stage while using Docker images.
**To Reproduce**
Steps to reproduce the behavior:
1. The command `docker run -it --rm -p 8080:8080 -t vdenotaris/spring-saml-sp:2.1.0-openjdk8` just fails when a new certificate is deployed for SSOCircle.
**Expected behavior**
Upgrade the certificate within the running stage instead the onbuild.
| non_infrastructure | usage of an external keystore describe the bug since ssocircle changes certificates quite often an external keystore is needed to properly separate the building stage and the running stage while using docker images to reproduce steps to reproduce the behavior the command docker run it rm p t vdenotaris spring saml sp just fails when a new certificate is deployed for ssocircle expected behavior upgrade the certificate within the running stage instead the onbuild | 0 |
25,294 | 18,437,612,378 | IssuesEvent | 2021-10-14 14:33:19 | spring-projects/spring-hateoas | https://api.github.com/repos/spring-projects/spring-hateoas | closed | Upgrade to Spring Framework 5.3.11 | process: in progress in: infrastructure type: task | To benefit from the fix for spring-projects/spring-framework#27439. Should fix #1448. | 1.0 | Upgrade to Spring Framework 5.3.11 - To benefit from the fix for spring-projects/spring-framework#27439. Should fix #1448. | infrastructure | upgrade to spring framework to benefit from the fix for spring projects spring framework should fix | 1 |
62,265 | 8,584,201,768 | IssuesEvent | 2018-11-13 22:00:35 | zalando-incubator/kube-ingress-aws-controller | https://api.github.com/repos/zalando-incubator/kube-ingress-aws-controller | closed | No ALB created when deploying a new ingress. | documentation | I am trying to setup skipper on aws with kops and I get all the way down to deploying my ingress, but it unfortunately doesn't get an ALB created for it.
Here are the details of my setup and exactly what I did.
```
kops version
Version 1.10.0
```
```
kubectl version
Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"bb9ffb1654d4a729bb4cec18ff088eacc153c239", GitTreeState:"clean", BuildDate:"2018-08-08T16:31:10Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.3", GitCommit:"2bba0127d85d5a46ab4b778548be28623b32d0b0", GitTreeState:"clean", BuildDate:"2018-05-21T09:05:37Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"}
```
I've set all the necessary IAM permissions on nodes through the spec of the cluster and I've added the required cloudLabel:
```yaml
apiVersion: kops/v1alpha2
kind: Cluster
...
spec:
...
cloudLabels:
kubernetes.io/cluster/<my cluster id>: owned
additionalPolicies:
node: |
[
{
"Effect": "Allow",
"Action": [
"autoscaling:DescribeAutoScalingGroups",
"autoscaling:AttachLoadBalancers",
"autoscaling:DetachLoadBalancers",
"autoscaling:DetachLoadBalancerTargetGroups",
"autoscaling:AttachLoadBalancerTargetGroups",
"autoscaling:DescribeLoadBalancerTargetGroups",
"elasticloadbalancing:DescribeLoadBalancers",
"elasticloadbalancing:CreateLoadBalancer",
"elasticloadbalancing:DeleteLoadBalancer",
"elasticloadbalancing:DescribeListeners",
"elasticloadbalancing:CreateListener",
"elasticloadbalancing:DeleteListener",
"elasticloadbalancing:DescribeTags",
"elasticloadbalancing:CreateTargetGroup",
"elasticloadbalancing:DeleteTargetGroup",
"elasticloadbalancing:DescribeTargetGroups",
"elasticloadbalancingv2:DescribeTargetGroups",
"elasticloadbalancingv2:DescribeLoadBalancers",
"elasticloadbalancingv2:CreateLoadBalancer",
"elasticloadbalancingv2:DeleteLoadBalancer",
"elasticloadbalancingv2:DescribeListeners",
"elasticloadbalancingv2:CreateListener",
"elasticloadbalancingv2:DeleteListener",
"elasticloadbalancingv2:DescribeTags",
"elasticloadbalancingv2:CreateTargetGroup",
"elasticloadbalancingv2:DeleteTargetGroup",
"ec2:DescribeInstances",
"ec2:DescribeSubnets",
"ec2:DescribeSecurityGroups",
"ec2:DescribeRouteTables",
"ec2:DescribeVpcs",
"acm:ListCertificates",
"acm:DescribeCertificate",
"iam:ListServerCertificates",
"iam:GetServerCertificate",
"cloudformation:Get*",
"cloudformation:Describe*",
"cloudformation:List*",
"cloudformation:Create*",
"cloudformation:Update*",
"cloudformation:Delete*"
],
"Resource": [
"*"
]
}
]
...
...
```
Created the security group through the cloudformation template with the proper VPC_ID and cluster name.
Deployed the `skipper.yaml`.
Deployed the `ingress-controller.yaml` with the proper region id.
Then the skipper-ingress pods started failing with:
```
level=error msg="failed to load all: request failed, status: 403, 403 Forbidden"
```
So I started searching for solutions and found this as a fix, which did resolve the problem and both skipper-ingress pods became healthy:
```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: ingress
rules:
- apiGroups:
- extensions
resources:
- ingresses
- ingresses/status
verbs:
- get
- list
- patch
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: skipper
rules:
- apiGroups:
- ""
resources:
- services
verbs:
- get
- list
# kubectl create clusterrolebinding skipper --clusterrole=ingress --serviceaccount=kube-system:default -n kube-system
---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
name: kube-system-default-user-role-binding-ingress
namespace: kube-system
subjects:
- kind: ServiceAccount
name: default
namespace: kube-system
roleRef:
kind: ClusterRole
name: ingress
apiGroup: rbac.authorization.k8s.io
# kubectl create clusterrolebinding skipper --clusterrole=skipper --serviceaccount=kube-system:default -n kube-system
---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
name: kube-system-default-user-role-binding-skipper
namespace: kube-system
subjects:
- kind: ServiceAccount
name: default
namespace: kube-system
roleRef:
kind: ClusterRole
name: skipper
apiGroup: rbac.authorization.k8s.io
```
And finally I have deployed my ingress, but it doesn't get an ALB (ADDRESS is empty).
Here it is:
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: name-of-ingress
labels:
app: app-name
chart: chart-name-0.1.0
release: release-name
heritage: Tiller
spec:
rules:
- host: example.com
http:
paths:
- backend:
serviceName: web-front
servicePort: 80
- host: api.example.com
http:
paths:
- path: /v0
backend:
serviceName: api-gateway
servicePort: 80
```
I checked the logs on the ingress controller and here is what I see once the ingress has been deployed in my cluster:
```
Start polling sleep 30s
Found 1 ingress(es)
Found 0 stack(s)
Found 1 auto scaling group(s)
Found 0 single instance(s)
Found 2 EC2 instance(s)
Found 2 certificate(s)
Have 1 model(s)
creating stack for certificates ["arn:aws:acm:<omitted>:certificate/<omitted>" "arn:aws:acm:<omitted>:certificate/<omitted>"] / ingress map["arn:aws:acm:<omitted>:certificate/<omitted>":["default/name-of-ingress"] "arn:aws:acm:<omitted>:certificate/<omitted>":["default/name-of-ingress"]]
stack "arn:aws:cloudformation:<omitted>:stack/kube-ingress-aws-controller-<my cluster id>-<omitted>/<omitted>" for certificates ["arn:aws:acm:<omitted>:certificate/<omitted>" "arn:aws:acm:<omitted>:certificate/<omitted>"] created
```
I don't see any errors on the ingress pods. What does "Found 0 stack(s)" and "Found 0 single instance(s)" mean? Is this the problem?
I have been battling with this for many hours now and I am completely out of ideas...
@szuecs what else can I provide for you to give you a better idea of what's going on, here? | 1.0 | No ALB created when deploying a new ingress. - I am trying to setup skipper on aws with kops and I get all the way down to deploying my ingress, but it unfortunately doesn't get an ALB created for it.
Here are the details of my setup and exactly what I did.
```
kops version
Version 1.10.0
```
```
kubectl version
Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"bb9ffb1654d4a729bb4cec18ff088eacc153c239", GitTreeState:"clean", BuildDate:"2018-08-08T16:31:10Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.3", GitCommit:"2bba0127d85d5a46ab4b778548be28623b32d0b0", GitTreeState:"clean", BuildDate:"2018-05-21T09:05:37Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"}
```
I've set all the necessary IAM permissions on nodes through the spec of the cluster and I've added the required cloudLabel:
```yaml
apiVersion: kops/v1alpha2
kind: Cluster
...
spec:
...
cloudLabels:
kubernetes.io/cluster/<my cluster id>: owned
additionalPolicies:
node: |
[
{
"Effect": "Allow",
"Action": [
"autoscaling:DescribeAutoScalingGroups",
"autoscaling:AttachLoadBalancers",
"autoscaling:DetachLoadBalancers",
"autoscaling:DetachLoadBalancerTargetGroups",
"autoscaling:AttachLoadBalancerTargetGroups",
"autoscaling:DescribeLoadBalancerTargetGroups",
"elasticloadbalancing:DescribeLoadBalancers",
"elasticloadbalancing:CreateLoadBalancer",
"elasticloadbalancing:DeleteLoadBalancer",
"elasticloadbalancing:DescribeListeners",
"elasticloadbalancing:CreateListener",
"elasticloadbalancing:DeleteListener",
"elasticloadbalancing:DescribeTags",
"elasticloadbalancing:CreateTargetGroup",
"elasticloadbalancing:DeleteTargetGroup",
"elasticloadbalancing:DescribeTargetGroups",
"elasticloadbalancingv2:DescribeTargetGroups",
"elasticloadbalancingv2:DescribeLoadBalancers",
"elasticloadbalancingv2:CreateLoadBalancer",
"elasticloadbalancingv2:DeleteLoadBalancer",
"elasticloadbalancingv2:DescribeListeners",
"elasticloadbalancingv2:CreateListener",
"elasticloadbalancingv2:DeleteListener",
"elasticloadbalancingv2:DescribeTags",
"elasticloadbalancingv2:CreateTargetGroup",
"elasticloadbalancingv2:DeleteTargetGroup",
"ec2:DescribeInstances",
"ec2:DescribeSubnets",
"ec2:DescribeSecurityGroups",
"ec2:DescribeRouteTables",
"ec2:DescribeVpcs",
"acm:ListCertificates",
"acm:DescribeCertificate",
"iam:ListServerCertificates",
"iam:GetServerCertificate",
"cloudformation:Get*",
"cloudformation:Describe*",
"cloudformation:List*",
"cloudformation:Create*",
"cloudformation:Update*",
"cloudformation:Delete*"
],
"Resource": [
"*"
]
}
]
...
...
```
Created the security group through the cloudformation template with the proper VPC_ID and cluster name.
Deployed the `skipper.yaml`.
Deployed the `ingress-controller.yaml` with the proper region id.
Then the skipper-ingress pods started failing with:
```
level=error msg="failed to load all: request failed, status: 403, 403 Forbidden"
```
So I started searching for solutions and found this as a fix, which did resolve the problem and both skipper-ingress pods became healthy:
```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: ingress
rules:
- apiGroups:
- extensions
resources:
- ingresses
- ingresses/status
verbs:
- get
- list
- patch
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: skipper
rules:
- apiGroups:
- ""
resources:
- services
verbs:
- get
- list
# kubectl create clusterrolebinding skipper --clusterrole=ingress --serviceaccount=kube-system:default -n kube-system
---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
name: kube-system-default-user-role-binding-ingress
namespace: kube-system
subjects:
- kind: ServiceAccount
name: default
namespace: kube-system
roleRef:
kind: ClusterRole
name: ingress
apiGroup: rbac.authorization.k8s.io
# kubectl create clusterrolebinding skipper --clusterrole=skipper --serviceaccount=kube-system:default -n kube-system
---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
name: kube-system-default-user-role-binding-skipper
namespace: kube-system
subjects:
- kind: ServiceAccount
name: default
namespace: kube-system
roleRef:
kind: ClusterRole
name: skipper
apiGroup: rbac.authorization.k8s.io
```
And finally I have deployed my ingress, but it doesn't get an ALB (ADDRESS is empty).
Here it is:
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: name-of-ingress
labels:
app: app-name
chart: chart-name-0.1.0
release: release-name
heritage: Tiller
spec:
rules:
- host: example.com
http:
paths:
- backend:
serviceName: web-front
servicePort: 80
- host: api.example.com
http:
paths:
- path: /v0
backend:
serviceName: api-gateway
servicePort: 80
```
I checked the logs on the ingress controller and here is what I see once the ingress has been deployed in my cluster:
```
Start polling sleep 30s
Found 1 ingress(es)
Found 0 stack(s)
Found 1 auto scaling group(s)
Found 0 single instance(s)
Found 2 EC2 instance(s)
Found 2 certificate(s)
Have 1 model(s)
creating stack for certificates ["arn:aws:acm:<omitted>:certificate/<omitted>" "arn:aws:acm:<omitted>:certificate/<omitted>"] / ingress map["arn:aws:acm:<omitted>:certificate/<omitted>":["default/name-of-ingress"] "arn:aws:acm:<omitted>:certificate/<omitted>":["default/name-of-ingress"]]
stack "arn:aws:cloudformation:<omitted>:stack/kube-ingress-aws-controller-<my cluster id>-<omitted>/<omitted>" for certificates ["arn:aws:acm:<omitted>:certificate/<omitted>" "arn:aws:acm:<omitted>:certificate/<omitted>"] created
```
I don't see any errors on the ingress pods. What does "Found 0 stack(s)" and "Found 0 single instance(s)" mean? Is this the problem?
I have been battling with this for many hours now and I am completely out of ideas...
@szuecs what else can I provide for you to give you a better idea of what's going on, here? | non_infrastructure | no alb created when deploying a new ingress i am trying to setup skipper on aws with kops and i get all the way down to deploying my ingress but it unfortunately doesn t get an alb created for it here are the details of my setup and exactly what i did kops version version kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform darwin server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux i ve set all the necessary iam permissions on nodes through the spec of the cluster and i ve added the required cloudlabel yaml apiversion kops kind cluster spec cloudlabels kubernetes io cluster owned additionalpolicies node effect allow action autoscaling describeautoscalinggroups autoscaling attachloadbalancers autoscaling detachloadbalancers autoscaling detachloadbalancertargetgroups autoscaling attachloadbalancertargetgroups autoscaling describeloadbalancertargetgroups elasticloadbalancing describeloadbalancers elasticloadbalancing createloadbalancer elasticloadbalancing deleteloadbalancer elasticloadbalancing describelisteners elasticloadbalancing createlistener elasticloadbalancing deletelistener elasticloadbalancing describetags elasticloadbalancing createtargetgroup elasticloadbalancing deletetargetgroup elasticloadbalancing describetargetgroups describetargetgroups describeloadbalancers createloadbalancer deleteloadbalancer describelisteners createlistener deletelistener describetags createtargetgroup deletetargetgroup describeinstances describesubnets describesecuritygroups describeroutetables describevpcs acm listcertificates acm describecertificate iam listservercertificates iam getservercertificate cloudformation get cloudformation describe cloudformation list cloudformation create cloudformation update cloudformation delete resource created the security group through the cloudformation template with the proper vpc id and cluster name deployed the skipper yaml deployed the ingress controller yaml with the proper region id then the skipper ingress pods started failing with level error msg failed to load all request failed status forbidden so i started searching for solutions and found this as a fix which did resolve the problem and both skipper ingress pods became healthy yaml apiversion rbac authorization io kind clusterrole metadata name ingress rules apigroups extensions resources ingresses ingresses status verbs get list patch apiversion rbac authorization io kind clusterrole metadata name skipper rules apigroups resources services verbs get list kubectl create clusterrolebinding skipper clusterrole ingress serviceaccount kube system default n kube system apiversion rbac authorization io kind clusterrolebinding metadata name kube system default user role binding ingress namespace kube system subjects kind serviceaccount name default namespace kube system roleref kind clusterrole name ingress apigroup rbac authorization io kubectl create clusterrolebinding skipper clusterrole skipper serviceaccount kube system default n kube system apiversion rbac authorization io kind clusterrolebinding metadata name kube system default user role binding skipper namespace kube system subjects kind serviceaccount name default namespace kube system roleref kind clusterrole name skipper apigroup rbac authorization io and finally i have deployed my ingress but it doesn t get an alb address is empty here it is yaml apiversion extensions kind ingress metadata name name of ingress labels app app name chart chart name release release name heritage tiller spec rules host example com http paths backend servicename web front serviceport host api example com http paths path backend servicename api gateway serviceport i checked the logs on the ingress controller and here is what i see once the ingress has been deployed in my cluster start polling sleep found ingress es found stack s found auto scaling group s found single instance s found instance s found certificate s have model s creating stack for certificates ingress map arn aws acm certificate stack arn aws cloudformation stack kube ingress aws controller for certificates created i don t see any errors on the ingress pods what does found stack s and found single instance s mean is this the problem i have been battling with this for many hours now and i am completely out of ideas szuecs what else can i provide for you to give you a better idea of what s going on here | 0 |
18,181 | 12,827,712,805 | IssuesEvent | 2020-07-06 19:02:38 | microsoft/react-native-windows | https://api.github.com/repos/microsoft/react-native-windows | closed | Add flowconfig updating functionality to override-tools or associated | Area: Infrastructure Area: Release Process Deforking nice-to-have | Facebook will often make upstream changes to their flowconfig file that we need to reflect in our own. We should automate this. | 1.0 | Add flowconfig updating functionality to override-tools or associated - Facebook will often make upstream changes to their flowconfig file that we need to reflect in our own. We should automate this. | infrastructure | add flowconfig updating functionality to override tools or associated facebook will often make upstream changes to their flowconfig file that we need to reflect in our own we should automate this | 1 |
52,451 | 27,569,090,244 | IssuesEvent | 2023-03-08 07:40:27 | hyperledger/besu | https://api.github.com/repos/hyperledger/besu | closed | Improve SLOAD performance | performance | <!-- Have you done the following? -->
<!-- * read the Code of Conduct? By filing an Issue, you are expected to -->
<!-- comply with it, including treating everyone with respect: -->
<!-- https://github.com/hyperledger/besu/blob/main/CODE_OF_CONDUCT.md -->
<!-- * Reproduced the issue in the latest version of the software -->
<!-- * Read the debugging docs: https://besu.hyperledger.org/en/stable/HowTo/Monitor/Logging/ -->
<!-- * Duplicate Issue check: https://github.com/search?q=+is%3Aissue+repo%3Ahyperledger/Besu -->
<!-- Note: Not all sections will apply to all issue types. -->
### Description
As a user, I want to have less IO on the database to keep performance open for other operations when using the getStorageValue method in UpdateTrackingAccount.
These calls are generated by SLOAD operation, we should make sure that [EIP-2929](https://eips.ethereum.org/EIPS/eip-2929) is correctly implemented in terms of caching accessed values.
### Acceptance Criteria
* The cache is hit only once if I need to read a storage value that is not being updated. I would like to use the cache for storage that is being updated as well but it is not as important.
This issue is to track an analysis for future performance improvements. This issue is not a bug but area of improvement. | True | Improve SLOAD performance - <!-- Have you done the following? -->
<!-- * read the Code of Conduct? By filing an Issue, you are expected to -->
<!-- comply with it, including treating everyone with respect: -->
<!-- https://github.com/hyperledger/besu/blob/main/CODE_OF_CONDUCT.md -->
<!-- * Reproduced the issue in the latest version of the software -->
<!-- * Read the debugging docs: https://besu.hyperledger.org/en/stable/HowTo/Monitor/Logging/ -->
<!-- * Duplicate Issue check: https://github.com/search?q=+is%3Aissue+repo%3Ahyperledger/Besu -->
<!-- Note: Not all sections will apply to all issue types. -->
### Description
As a user, I want to have less IO on the database to keep performance open for other operations when using the getStorageValue method in UpdateTrackingAccount.
These calls are generated by SLOAD operation, we should make sure that [EIP-2929](https://eips.ethereum.org/EIPS/eip-2929) is correctly implemented in terms of caching accessed values.
### Acceptance Criteria
* The cache is hit only once if I need to read a storage value that is not being updated. I would like to use the cache for storage that is being updated as well but it is not as important.
This issue is to track an analysis for future performance improvements. This issue is not a bug but area of improvement. | non_infrastructure | improve sload performance description as a user i want to have less io on the database to keep performance open for other operations when using the getstoragevalue method in updatetrackingaccount these calls are generated by sload operation we should make sure that is correctly implemented in terms of caching accessed values acceptance criteria the cache is hit only once if i need to read a storage value that is not being updated i would like to use the cache for storage that is being updated as well but it is not as important this issue is to track an analysis for future performance improvements this issue is not a bug but area of improvement | 0 |
32,270 | 26,587,439,218 | IssuesEvent | 2023-01-23 03:53:19 | zer0Kerbal/SeaDragon | https://api.github.com/repos/zer0Kerbal/SeaDragon | closed | Update License | issue: license type: infrastructure | # Update License
<!--
updateLicense v1.0.0.0
created: 17 Aug 2022
updated:
-->
* Update License
* [ ] Updated License: {newLicense}
* was: {CC BY-NC-SA 4.0}
this file: This file: All Rights Reserved by zer0Kerbal
| 1.0 | Update License - # Update License
<!--
updateLicense v1.0.0.0
created: 17 Aug 2022
updated:
-->
* Update License
* [ ] Updated License: {newLicense}
* was: {CC BY-NC-SA 4.0}
this file: This file: All Rights Reserved by zer0Kerbal
| infrastructure | update license update license updatelicense created aug updated update license updated license newlicense was cc by nc sa this file this file all rights reserved by | 1 |
35,017 | 14,577,901,092 | IssuesEvent | 2020-12-18 03:18:13 | invertase/react-native-firebase | https://api.github.com/repos/invertase/react-native-firebase | closed | [Crash] putFile will crash when extension in unknown | Impact: Crash Platform: iOS Priority: P1 Service: Storage Type: Invertase Workflow: Waiting for User Response | I'm running xcode12 on Apple Silicon and it seems that pathExtension which used to return the extensions of the file in path is returning nil on iOS14 simulators (for gif files - does not happen on iOS13). This is probably a simulator issue, however, this code does not check if UTI is null and therefor crashes when trying to release:
`CFRelease(UTI);`
https://github.com/invertase/react-native-firebase/blob/b5357573e97ccc3bfdf6a190713a7df66e462c4b/packages/storage/ios/RNFBStorage/RNFBStorageCommon.m#L62-L66
It should be something like:
`if (UTI) {
CFRelease(UTI);
}`
Return value can be null according to the [documentation](https://developer.apple.com/documentation/coreservices/1448939-uttypecreatepreferredidentifierf):
> Return Value
> A new CFStringRef containing a uniform type identifier, or NULL if inTagClass is not a known tag class | 1.0 | [Crash] putFile will crash when extension in unknown - I'm running xcode12 on Apple Silicon and it seems that pathExtension which used to return the extensions of the file in path is returning nil on iOS14 simulators (for gif files - does not happen on iOS13). This is probably a simulator issue, however, this code does not check if UTI is null and therefor crashes when trying to release:
`CFRelease(UTI);`
https://github.com/invertase/react-native-firebase/blob/b5357573e97ccc3bfdf6a190713a7df66e462c4b/packages/storage/ios/RNFBStorage/RNFBStorageCommon.m#L62-L66
It should be something like:
`if (UTI) {
CFRelease(UTI);
}`
Return value can be null according to the [documentation](https://developer.apple.com/documentation/coreservices/1448939-uttypecreatepreferredidentifierf):
> Return Value
> A new CFStringRef containing a uniform type identifier, or NULL if inTagClass is not a known tag class | non_infrastructure | putfile will crash when extension in unknown i m running on apple silicon and it seems that pathextension which used to return the extensions of the file in path is returning nil on simulators for gif files does not happen on this is probably a simulator issue however this code does not check if uti is null and therefor crashes when trying to release cfrelease uti it should be something like if uti cfrelease uti return value can be null according to the return value a new cfstringref containing a uniform type identifier or null if intagclass is not a known tag class | 0 |
4,419 | 3,023,733,683 | IssuesEvent | 2015-08-01 20:30:47 | ash-lang/ash | https://api.github.com/repos/ash-lang/ash | closed | Definition files | analysis code-gen grammar | Definition files will be files that define global functions, operators and import declarations. Definition files will have the extension `.ashd` (the `d` is short for "definition") and will consist of (all of which are optional) a package declaration, import declarations, operator overloads, operator definitions and extension functions. Definition files will be "included" into the current ash source file with the `include` keyword. | 1.0 | Definition files - Definition files will be files that define global functions, operators and import declarations. Definition files will have the extension `.ashd` (the `d` is short for "definition") and will consist of (all of which are optional) a package declaration, import declarations, operator overloads, operator definitions and extension functions. Definition files will be "included" into the current ash source file with the `include` keyword. | non_infrastructure | definition files definition files will be files that define global functions operators and import declarations definition files will have the extension ashd the d is short for definition and will consist of all of which are optional a package declaration import declarations operator overloads operator definitions and extension functions definition files will be included into the current ash source file with the include keyword | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.