Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
237,149 | 26,078,820,352 | IssuesEvent | 2022-12-25 01:20:44 | LibrIT/passhport | https://api.github.com/repos/LibrIT/passhport | opened | CVE-2022-40898 (Medium) detected in wheel-0.37.1-py2.py3-none-any.whl | security vulnerability | ## CVE-2022-40898 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>wheel-0.37.1-py2.py3-none-any.whl</b></p></summary>
<p>A built-package format for Python</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/27/d6/003e593296a85fd6ed616ed962795b2f87709c3eee2bca4f6d0fe55c6d00/wheel-0.37.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/27/d6/003e593296a85fd6ed616ed962795b2f87709c3eee2bca4f6d0fe55c6d00/wheel-0.37.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **wheel-0.37.1-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LibrIT/passhport/commit/85f9855d31f91d439049909da1bfb4711986a3da">85f9855d31f91d439049909da1bfb4711986a3da</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue discovered in Python Packaging Authority (PyPA) Wheel 0.37.1 and earlier allows remote attackers to cause a denial of service via attacker controlled input to wheel cli.
<p>Publish Date: 2022-12-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40898>CVE-2022-40898</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-12-23</p>
<p>Fix Resolution: wheel 0.38.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-40898 (Medium) detected in wheel-0.37.1-py2.py3-none-any.whl - ## CVE-2022-40898 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>wheel-0.37.1-py2.py3-none-any.whl</b></p></summary>
<p>A built-package format for Python</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/27/d6/003e593296a85fd6ed616ed962795b2f87709c3eee2bca4f6d0fe55c6d00/wheel-0.37.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/27/d6/003e593296a85fd6ed616ed962795b2f87709c3eee2bca4f6d0fe55c6d00/wheel-0.37.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **wheel-0.37.1-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LibrIT/passhport/commit/85f9855d31f91d439049909da1bfb4711986a3da">85f9855d31f91d439049909da1bfb4711986a3da</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue discovered in Python Packaging Authority (PyPA) Wheel 0.37.1 and earlier allows remote attackers to cause a denial of service via attacker controlled input to wheel cli.
<p>Publish Date: 2022-12-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40898>CVE-2022-40898</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-12-23</p>
<p>Fix Resolution: wheel 0.38.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in wheel none any whl cve medium severity vulnerability vulnerable library wheel none any whl a built package format for python library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x wheel none any whl vulnerable library found in head commit a href found in base branch master vulnerability details an issue discovered in python packaging authority pypa wheel and earlier allows remote attackers to cause a denial of service via attacker controlled input to wheel cli publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution wheel step up your open source security game with mend | 0 |
929 | 2,672,654,935 | IssuesEvent | 2015-03-24 15:17:46 | cobbler/cobbler | https://api.github.com/repos/cobbler/cobbler | closed | Static Routes field in gui can break the setting silently | feature_request usability WEB | I have a long list of nets and their routes, and I pasted in them in the web gui. But, I had cut them from a document with quotes, like this "net:route net:route" and I happened to include the quotes.
As I pressed SAVE I realized that it might not work, but when I then went back and looked at it, the list looked quite ok, without the quotes.
When I installed the machine the routes where broken and it was only when I ran "cobbler system report --name=FOO" that it became clear that the Static Route list was not individual nets and routes, quoted and comma separated, but a long list within quotes.
So, the GUI accepted wrong input and did not complain, nor did it split the string up into the constituent strings and comma separated them.
Considering https://github.com/cobbler/cobbler/issues/291 did not fix the problem on the slave after the replication master had been fixed, this is a most aggravating behaviour. | True | Static Routes field in gui can break the setting silently - I have a long list of nets and their routes, and I pasted in them in the web gui. But, I had cut them from a document with quotes, like this "net:route net:route" and I happened to include the quotes.
As I pressed SAVE I realized that it might not work, but when I then went back and looked at it, the list looked quite ok, without the quotes.
When I installed the machine the routes where broken and it was only when I ran "cobbler system report --name=FOO" that it became clear that the Static Route list was not individual nets and routes, quoted and comma separated, but a long list within quotes.
So, the GUI accepted wrong input and did not complain, nor did it split the string up into the constituent strings and comma separated them.
Considering https://github.com/cobbler/cobbler/issues/291 did not fix the problem on the slave after the replication master had been fixed, this is a most aggravating behaviour. | non_priority | static routes field in gui can break the setting silently i have a long list of nets and their routes and i pasted in them in the web gui but i had cut them from a document with quotes like this net route net route and i happened to include the quotes as i pressed save i realized that it might not work but when i then went back and looked at it the list looked quite ok without the quotes when i installed the machine the routes where broken and it was only when i ran cobbler system report name foo that it became clear that the static route list was not individual nets and routes quoted and comma separated but a long list within quotes so the gui accepted wrong input and did not complain nor did it split the string up into the constituent strings and comma separated them considering did not fix the problem on the slave after the replication master had been fixed this is a most aggravating behaviour | 0 |
557,993 | 16,523,888,207 | IssuesEvent | 2021-05-26 17:28:27 | TryGhost/Ghost | https://api.github.com/repos/TryGhost/Ghost | closed | The price change is not updated for the helper {{price @price.monthly}} and {{price @price.yearly}} | bug members / mega p1 - priority themes / frontend | Hi!
In the Ghost 4.6.0 the price change is not updated for custom "membership" page ('Portal' works fine).
### To Reproduce
1. Install Ghost with a new database.
2. Connect with Stripe (the default prices of $5/month and $50/year will be set).
3. After a price change, the custom page is not updated:
Dashboard:

Theme:

Stripe:

When you click through to Stripe checkout, the price is properly set.
### Technical details:
* Ghost Version: 4.6.0
* Node Version: 12.22.1
* Browser/OS: All
* Database: sqlite3
| 1.0 | The price change is not updated for the helper {{price @price.monthly}} and {{price @price.yearly}} - Hi!
In the Ghost 4.6.0 the price change is not updated for custom "membership" page ('Portal' works fine).
### To Reproduce
1. Install Ghost with a new database.
2. Connect with Stripe (the default prices of $5/month and $50/year will be set).
3. After a price change, the custom page is not updated:
Dashboard:

Theme:

Stripe:

When you click through to Stripe checkout, the price is properly set.
### Technical details:
* Ghost Version: 4.6.0
* Node Version: 12.22.1
* Browser/OS: All
* Database: sqlite3
| priority | the price change is not updated for the helper price price monthly and price price yearly hi in the ghost the price change is not updated for custom membership page portal works fine to reproduce install ghost with a new database connect with stripe the default prices of month and year will be set after a price change the custom page is not updated dashboard theme stripe when you click through to stripe checkout the price is properly set technical details ghost version node version browser os all database | 1 |
19,585 | 2,622,154,524 | IssuesEvent | 2015-03-04 00:07:32 | byzhang/terrastore | https://api.github.com/repos/byzhang/terrastore | closed | Implement conditional put operation | auto-migrated Milestone-0.5.0 Priority-High Project-Terrastore Type-Feature | ```
Terrastore should support conditional put operation, in order to provide
atomic "compare-and-swap" semantics (very useful for optimistic concurrency
control and alike).
```
Original issue reported on code.google.com by `sergio.b...@gmail.com` on 4 Mar 2010 at 6:39
* Blocked on: #37 | 1.0 | Implement conditional put operation - ```
Terrastore should support conditional put operation, in order to provide
atomic "compare-and-swap" semantics (very useful for optimistic concurrency
control and alike).
```
Original issue reported on code.google.com by `sergio.b...@gmail.com` on 4 Mar 2010 at 6:39
* Blocked on: #37 | priority | implement conditional put operation terrastore should support conditional put operation in order to provide atomic compare and swap semantics very useful for optimistic concurrency control and alike original issue reported on code google com by sergio b gmail com on mar at blocked on | 1 |
757,600 | 26,520,456,760 | IssuesEvent | 2023-01-19 01:48:37 | kubernetes/website | https://api.github.com/repos/kubernetes/website | closed | Create a Tasks doc for Service Account Issuer Discovery | kind/feature priority/backlog lifecycle/frozen | **This is a Feature Request**
**What would you like to be added**
Add a Tasks doc for configuring Service Account Issuer Discovery before the feature goes to beta.
**Why is this needed**
To help users understand how to configure the feature correctly.
**Comments**
@sftim recommended we make this a beta requirement https://github.com/kubernetes/website/pull/19328#pullrequestreview-365664336
/assign @mtaufen
/cc @sftim @mikedanese
| 1.0 | Create a Tasks doc for Service Account Issuer Discovery - **This is a Feature Request**
**What would you like to be added**
Add a Tasks doc for configuring Service Account Issuer Discovery before the feature goes to beta.
**Why is this needed**
To help users understand how to configure the feature correctly.
**Comments**
@sftim recommended we make this a beta requirement https://github.com/kubernetes/website/pull/19328#pullrequestreview-365664336
/assign @mtaufen
/cc @sftim @mikedanese
| priority | create a tasks doc for service account issuer discovery this is a feature request what would you like to be added add a tasks doc for configuring service account issuer discovery before the feature goes to beta why is this needed to help users understand how to configure the feature correctly comments sftim recommended we make this a beta requirement assign mtaufen cc sftim mikedanese | 1 |
159,548 | 20,068,342,946 | IssuesEvent | 2022-02-04 01:13:55 | harrinry/core | https://api.github.com/repos/harrinry/core | opened | CVE-2020-28469 (High) detected in glob-parent-3.1.0.tgz | security vulnerability | ## CVE-2020-28469 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- ckeditor5-dev-utils-25.4.5.tgz (Root Library)
- postcss-mixins-6.2.3.tgz
- globby-8.0.2.tgz
- fast-glob-2.2.7.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/harrinry/core/commit/b709d5d52c63939480cdc94350a24a9aca4df838">b709d5d52c63939480cdc94350a24a9aca4df838</a></p>
<p>Found in base branch: <b>9.4.x</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"3.1.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@ckeditor/ckeditor5-dev-utils:25.4.5;postcss-mixins:6.2.3;globby:8.0.2;fast-glob:2.2.7;glob-parent:3.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2","isBinary":false}],"baseBranches":["9.4.x"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-28469 (High) detected in glob-parent-3.1.0.tgz - ## CVE-2020-28469 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- ckeditor5-dev-utils-25.4.5.tgz (Root Library)
- postcss-mixins-6.2.3.tgz
- globby-8.0.2.tgz
- fast-glob-2.2.7.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/harrinry/core/commit/b709d5d52c63939480cdc94350a24a9aca4df838">b709d5d52c63939480cdc94350a24a9aca4df838</a></p>
<p>Found in base branch: <b>9.4.x</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"3.1.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@ckeditor/ckeditor5-dev-utils:25.4.5;postcss-mixins:6.2.3;globby:8.0.2;fast-glob:2.2.7;glob-parent:3.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2","isBinary":false}],"baseBranches":["9.4.x"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in glob parent tgz cve high severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy dev utils tgz root library postcss mixins tgz globby tgz fast glob tgz x glob parent tgz vulnerable library found in head commit a href found in base branch x vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree ckeditor dev utils postcss mixins globby fast glob glob parent isminimumfixversionavailable true minimumfixversion glob parent isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator vulnerabilityurl | 0 |
47,710 | 13,248,508,716 | IssuesEvent | 2020-08-19 19:05:57 | kenferrara/cbp-theme | https://api.github.com/repos/kenferrara/cbp-theme | opened | CVE-2018-11695 (High) detected in node-sass-4.11.0.tgz | security vulnerability | ## CVE-2018-11695 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/cbp-theme/cbp-theme/package.json</p>
<p>Path to vulnerable library: /cbp-theme/cbp-theme/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/cbp-theme/commit/00f1482f5efa0120a277f069fffcee0de8e6adec">00f1482f5efa0120a277f069fffcee0de8e6adec</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass <3.5.3. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11695>CVE-2018-11695</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/issues/2664">https://github.com/sass/libsass/issues/2664</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: Libsass:3.5.3, Node-sass:4.9.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"4.11.0","isTransitiveDependency":false,"dependencyTree":"node-sass:4.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Libsass:3.5.3, Node-sass:4.9.0"}],"vulnerabilityIdentifier":"CVE-2018-11695","vulnerabilityDetails":"An issue was discovered in LibSass \u003c3.5.3. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11695","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-11695 (High) detected in node-sass-4.11.0.tgz - ## CVE-2018-11695 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/cbp-theme/cbp-theme/package.json</p>
<p>Path to vulnerable library: /cbp-theme/cbp-theme/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/cbp-theme/commit/00f1482f5efa0120a277f069fffcee0de8e6adec">00f1482f5efa0120a277f069fffcee0de8e6adec</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass <3.5.3. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11695>CVE-2018-11695</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/issues/2664">https://github.com/sass/libsass/issues/2664</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: Libsass:3.5.3, Node-sass:4.9.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"4.11.0","isTransitiveDependency":false,"dependencyTree":"node-sass:4.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Libsass:3.5.3, Node-sass:4.9.0"}],"vulnerabilityIdentifier":"CVE-2018-11695","vulnerabilityDetails":"An issue was discovered in LibSass \u003c3.5.3. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11695","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in node sass tgz cve high severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm cbp theme cbp theme package json path to vulnerable library cbp theme cbp theme node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details an issue was discovered in libsass a null pointer dereference was found in the function sass expand operator which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in libsass a null pointer dereference was found in the function sass expand operator which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact vulnerabilityurl | 0 |
45,458 | 13,123,261,707 | IssuesEvent | 2020-08-06 00:00:29 | jtimberlake/paypal-checkout-components | https://api.github.com/repos/jtimberlake/paypal-checkout-components | opened | CVE-2018-3737 (High) detected in sshpk-1.13.1.tgz | security vulnerability | ## CVE-2018-3737 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sshpk-1.13.1.tgz</b></p></summary>
<p>A library for finding and using SSH public keys</p>
<p>Library home page: <a href="https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz">https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/paypal-checkout-components/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/paypal-checkout-components/node_modules/npx/node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/sshpk/package.json</p>
<p>
Dependency Hierarchy:
- grumbler-scripts-3.0.76.tgz (Root Library)
- npx-10.2.2.tgz
- npm-5.1.0.tgz
- request-2.81.0.tgz
- http-signature-1.1.1.tgz
- :x: **sshpk-1.13.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/paypal-checkout-components/commit/a697d69fc10f0ab278e7bd0b384bd96bbdda9216">a697d69fc10f0ab278e7bd0b384bd96bbdda9216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
sshpk is vulnerable to ReDoS when parsing crafted invalid public keys.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737>CVE-2018-3737</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/319593">https://hackerone.com/reports/319593</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.13.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"sshpk","packageVersion":"1.13.1","isTransitiveDependency":true,"dependencyTree":"grumbler-scripts:3.0.76;npx:10.2.2;npm:5.1.0;request:2.81.0;http-signature:1.1.1;sshpk:1.13.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13.2"}],"vulnerabilityIdentifier":"CVE-2018-3737","vulnerabilityDetails":"sshpk is vulnerable to ReDoS when parsing crafted invalid public keys.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-3737 (High) detected in sshpk-1.13.1.tgz - ## CVE-2018-3737 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sshpk-1.13.1.tgz</b></p></summary>
<p>A library for finding and using SSH public keys</p>
<p>Library home page: <a href="https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz">https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/paypal-checkout-components/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/paypal-checkout-components/node_modules/npx/node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/sshpk/package.json</p>
<p>
Dependency Hierarchy:
- grumbler-scripts-3.0.76.tgz (Root Library)
- npx-10.2.2.tgz
- npm-5.1.0.tgz
- request-2.81.0.tgz
- http-signature-1.1.1.tgz
- :x: **sshpk-1.13.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/paypal-checkout-components/commit/a697d69fc10f0ab278e7bd0b384bd96bbdda9216">a697d69fc10f0ab278e7bd0b384bd96bbdda9216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
sshpk is vulnerable to ReDoS when parsing crafted invalid public keys.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737>CVE-2018-3737</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/319593">https://hackerone.com/reports/319593</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.13.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"sshpk","packageVersion":"1.13.1","isTransitiveDependency":true,"dependencyTree":"grumbler-scripts:3.0.76;npx:10.2.2;npm:5.1.0;request:2.81.0;http-signature:1.1.1;sshpk:1.13.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13.2"}],"vulnerabilityIdentifier":"CVE-2018-3737","vulnerabilityDetails":"sshpk is vulnerable to ReDoS when parsing crafted invalid public keys.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in sshpk tgz cve high severity vulnerability vulnerable library sshpk tgz a library for finding and using ssh public keys library home page a href path to dependency file tmp ws scm paypal checkout components package json path to vulnerable library tmp ws scm paypal checkout components node modules npx node modules npm node modules request node modules http signature node modules sshpk package json dependency hierarchy grumbler scripts tgz root library npx tgz npm tgz request tgz http signature tgz x sshpk tgz vulnerable library found in head commit a href vulnerability details sshpk is vulnerable to redos when parsing crafted invalid public keys publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails sshpk is vulnerable to redos when parsing crafted invalid public keys vulnerabilityurl | 0 |
217,296 | 24,325,987,884 | IssuesEvent | 2022-09-30 14:49:11 | Sharecare/capacitor-plugin-aep-analytics | https://api.github.com/repos/Sharecare/capacitor-plugin-aep-analytics | opened | CVE-2021-36090 (High) detected in commons-compress-1.12.jar | security vulnerability | ## CVE-2021-36090 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.12.jar</b></p></summary>
<p>Apache Commons Compress software defines an API for working with
compression and archive formats. These include: bzip2, gzip, pack200,
lzma, xz, Snappy, traditional Unix Compress, DEFLATE and ar, cpio,
jar, tar, zip, dump, 7z, arj.</p>
<p>Path to dependency file: /android/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.12/84caa68576e345eb5e7ae61a0e5a9229eb100d7b/commons-compress-1.12.jar</p>
<p>
Dependency Hierarchy:
- lint-gradle-27.2.1.jar (Root Library)
- builder-4.2.1.jar
- sdklib-27.2.1.jar
- :x: **commons-compress-1.12.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Sharecare/capacitor-plugin-aep-analytics/commit/359199a86cd4095312b4582afe7ad22e605dca6f">359199a86cd4095312b4582afe7ad22e605dca6f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When reading a specially crafted ZIP archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' zip package.
<p>Publish Date: 2021-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-36090>CVE-2021-36090</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p>
<p>Release Date: 2021-07-13</p>
<p>Fix Resolution: org.apache.commons:commons-compress:1.21</p>
</p>
</details>
<p></p>
| True | CVE-2021-36090 (High) detected in commons-compress-1.12.jar - ## CVE-2021-36090 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.12.jar</b></p></summary>
<p>Apache Commons Compress software defines an API for working with
compression and archive formats. These include: bzip2, gzip, pack200,
lzma, xz, Snappy, traditional Unix Compress, DEFLATE and ar, cpio,
jar, tar, zip, dump, 7z, arj.</p>
<p>Path to dependency file: /android/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.12/84caa68576e345eb5e7ae61a0e5a9229eb100d7b/commons-compress-1.12.jar</p>
<p>
Dependency Hierarchy:
- lint-gradle-27.2.1.jar (Root Library)
- builder-4.2.1.jar
- sdklib-27.2.1.jar
- :x: **commons-compress-1.12.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Sharecare/capacitor-plugin-aep-analytics/commit/359199a86cd4095312b4582afe7ad22e605dca6f">359199a86cd4095312b4582afe7ad22e605dca6f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When reading a specially crafted ZIP archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' zip package.
<p>Publish Date: 2021-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-36090>CVE-2021-36090</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p>
<p>Release Date: 2021-07-13</p>
<p>Fix Resolution: org.apache.commons:commons-compress:1.21</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in commons compress jar cve high severity vulnerability vulnerable library commons compress jar apache commons compress software defines an api for working with compression and archive formats these include gzip lzma xz snappy traditional unix compress deflate and ar cpio jar tar zip dump arj path to dependency file android build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons compress commons compress jar dependency hierarchy lint gradle jar root library builder jar sdklib jar x commons compress jar vulnerable library found in head commit a href found in base branch main vulnerability details when reading a specially crafted zip archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress zip package publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache commons commons compress | 0 |
84,323 | 3,663,308,813 | IssuesEvent | 2016-02-19 04:53:00 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Additional image tiles on the create page should display their individual logo icon | area/usability component/web kind/enhancement priority/P2 | We have most, if not all, logos for these in the icon-font set. Currently the generic cube icon is being used.
<img width="854" alt="missing-logos" src="https://cloud.githubusercontent.com/assets/1874151/11127790/fe45e306-8943-11e5-86de-72999e941602.png">
| 1.0 | Additional image tiles on the create page should display their individual logo icon - We have most, if not all, logos for these in the icon-font set. Currently the generic cube icon is being used.
<img width="854" alt="missing-logos" src="https://cloud.githubusercontent.com/assets/1874151/11127790/fe45e306-8943-11e5-86de-72999e941602.png">
| priority | additional image tiles on the create page should display their individual logo icon we have most if not all logos for these in the icon font set currently the generic cube icon is being used img width alt missing logos src | 1 |
6,857 | 9,167,689,179 | IssuesEvent | 2019-03-02 15:55:00 | JurassiCraftTeam/JurassiCraft2 | https://api.github.com/repos/JurassiCraftTeam/JurassiCraft2 | closed | Crash with SpongeForge and JurassiCraft | 1.12.2 crash incompatibility missing logs more info needed needs verification | I used forge-2778.sponge server
When I put Jurassic mod into the server, it will prompt level.dat to be modified. If I click ok, a lot of overworld's original items will be replaced with all sorts of weird stuff, which is very confusing.The following warning will be displayed on the server:
[11:56:41] [Server thread/ERROR] [minecraft/SaveFormatOld]: Exception reading .\world\level.dat
java.lang.RuntimeException: Invalid id 4096 - maximum id range exceeded.
at net.minecraftforge.registries.ForgeRegistry.add(ForgeRegistry.java:297) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at net.minecraftforge.registries.ForgeRegistry.loadIds(ForgeRegistry.java:695) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at net.minecraftforge.registries.GameData.loadFrozenDataToStagingRegistry(GameData.java:750) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at net.minecraftforge.registries.GameData.lambda$injectSnapshot$25(GameData.java:693) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at com.google.common.collect.HashBiMap.forEach(HashBiMap.java:528) ~[min
ecraft_server.1.12.2.jar:?]
at net.minecraftforge.registries.GameData.injectSnapshot(GameData.java:690) ~[GameData.class:?]
at net.minecraftforge.fml.common.FMLContainer.readData(FMLContainer.java:172) ~[FMLContainer.class:?]
at net.minecraftforge.fml.common.FMLCommonHandler.handleWorldDataLoad(FMLCommonHandler.java:423) ~[FMLCommonHandler.class:?]
at net.minecraft.world.storage.SaveFormatOld.loadAndFix(SaveFormatOld.java:119) [bfc.class:?] at net.minecraft.world.storage.SaveHandler.redirect$onLoadWorldInfo$bcl0
00(SaveHandler.java:1068) [bfb.class:?]
at net.minecraft.world.storage.SaveHandler.func_75757_d(SaveHandler.java:122) [bfb.class:?]
at org.spongepowered.common.world.WorldManager.loadAllWorlds(WorldManager.java:711) [WorldManager.class:1.12.2-2768-7.1.4]
at net.minecraft.server.MinecraftServer.func_71247_a(MinecraftServer.java:3545) [MinecraftServer.class:?]
at net.minecraft.server.dedicated.DedicatedServer.func_71197_b(DedicatedServer.java:270) [nz.class:?]
at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:486) [MinecraftServer.class:?]
at java.lang.Thread.run(Unknown Source) [?:1.8.0_161] | True | Crash with SpongeForge and JurassiCraft - I used forge-2778.sponge server
When I put Jurassic mod into the server, it will prompt level.dat to be modified. If I click ok, a lot of overworld's original items will be replaced with all sorts of weird stuff, which is very confusing.The following warning will be displayed on the server:
[11:56:41] [Server thread/ERROR] [minecraft/SaveFormatOld]: Exception reading .\world\level.dat
java.lang.RuntimeException: Invalid id 4096 - maximum id range exceeded.
at net.minecraftforge.registries.ForgeRegistry.add(ForgeRegistry.java:297) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at net.minecraftforge.registries.ForgeRegistry.loadIds(ForgeRegistry.java:695) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at net.minecraftforge.registries.GameData.loadFrozenDataToStagingRegistry(GameData.java:750) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at net.minecraftforge.registries.GameData.lambda$injectSnapshot$25(GameData.java:693) ~[forge-1.12.2-14.23.5.2778-universal.jar:?]
at com.google.common.collect.HashBiMap.forEach(HashBiMap.java:528) ~[min
ecraft_server.1.12.2.jar:?]
at net.minecraftforge.registries.GameData.injectSnapshot(GameData.java:690) ~[GameData.class:?]
at net.minecraftforge.fml.common.FMLContainer.readData(FMLContainer.java:172) ~[FMLContainer.class:?]
at net.minecraftforge.fml.common.FMLCommonHandler.handleWorldDataLoad(FMLCommonHandler.java:423) ~[FMLCommonHandler.class:?]
at net.minecraft.world.storage.SaveFormatOld.loadAndFix(SaveFormatOld.java:119) [bfc.class:?] at net.minecraft.world.storage.SaveHandler.redirect$onLoadWorldInfo$bcl0
00(SaveHandler.java:1068) [bfb.class:?]
at net.minecraft.world.storage.SaveHandler.func_75757_d(SaveHandler.java:122) [bfb.class:?]
at org.spongepowered.common.world.WorldManager.loadAllWorlds(WorldManager.java:711) [WorldManager.class:1.12.2-2768-7.1.4]
at net.minecraft.server.MinecraftServer.func_71247_a(MinecraftServer.java:3545) [MinecraftServer.class:?]
at net.minecraft.server.dedicated.DedicatedServer.func_71197_b(DedicatedServer.java:270) [nz.class:?]
at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:486) [MinecraftServer.class:?]
at java.lang.Thread.run(Unknown Source) [?:1.8.0_161] | non_priority | crash with spongeforge and jurassicraft i used forge sponge server when i put jurassic mod into the server it will prompt level dat to be modified if i click ok a lot of overworld s original items will be replaced with all sorts of weird stuff which is very confusing the following warning will be displayed on the server exception reading world level dat java lang runtimeexception invalid id maximum id range exceeded at net minecraftforge registries forgeregistry add forgeregistry java at net minecraftforge registries forgeregistry loadids forgeregistry java at net minecraftforge registries gamedata loadfrozendatatostagingregistry gamedata java at net minecraftforge registries gamedata lambda injectsnapshot gamedata java at com google common collect hashbimap foreach hashbimap java min ecraft server jar at net minecraftforge registries gamedata injectsnapshot gamedata java at net minecraftforge fml common fmlcontainer readdata fmlcontainer java at net minecraftforge fml common fmlcommonhandler handleworlddataload fmlcommonhandler java at net minecraft world storage saveformatold loadandfix saveformatold java at net minecraft world storage savehandler redirect onloadworldinfo savehandler java at net minecraft world storage savehandler func d savehandler java at org spongepowered common world worldmanager loadallworlds worldmanager java at net minecraft server minecraftserver func a minecraftserver java at net minecraft server dedicated dedicatedserver func b dedicatedserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run unknown source | 0 |
21,254 | 6,132,543,974 | IssuesEvent | 2017-06-25 03:35:29 | ganeti/ganeti | https://api.github.com/repos/ganeti/ganeti | closed | Speed up build time | imported_from_google_code Status:Obsolete Type-Refactoring | Originally reported of Google Code with ID 360.
```
It would be nice if we could build each haskell file just once, if possible.
Thanks,
Guido
```
Originally added on 2013-02-06 15:03:47 +0000 UTC. | 1.0 | Speed up build time - Originally reported of Google Code with ID 360.
```
It would be nice if we could build each haskell file just once, if possible.
Thanks,
Guido
```
Originally added on 2013-02-06 15:03:47 +0000 UTC. | non_priority | speed up build time originally reported of google code with id it would be nice if we could build each haskell file just once if possible thanks guido originally added on utc | 0 |
294,598 | 25,386,593,292 | IssuesEvent | 2022-11-21 22:30:46 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [APM] Flyout uses full screen | bug Team:APM apm:test-plan-regression | ### Issue
https://user-images.githubusercontent.com/3369346/203170812-e91a736d-09f1-42e5-b9b5-1bae84306992.mov
<img width="1791" alt="image" src="https://user-images.githubusercontent.com/3369346/203170951-d0e04645-3ab5-4ab6-858d-c6ea1a1c346b.png">
Some of the places we use `Flyout` component
```
apm • public/components/app/settings/agent_keys/create_agent_key.tsx:
99 return (
100: <EuiFlyout onClose={onCancel} size="s">
101 <EuiFlyoutHeader hasBorder>
apm • public/components/app/settings/custom_link/create_edit_custom_link_flyout/index.tsx:
78 <form onSubmit={onSubmit} id="customLink_form">
79: <EuiFlyout ownFocus onClose={onClose} size="m">
80 <EuiFlyoutHeader hasBorder>
apm • public/components/shared/apm_header_action_menu/labs/labs_flyout.tsx:
85 return (
86: <EuiFlyout onClose={onClose}>
87 <EuiFlyoutHeader hasBorder>
``` | 1.0 | [APM] Flyout uses full screen - ### Issue
https://user-images.githubusercontent.com/3369346/203170812-e91a736d-09f1-42e5-b9b5-1bae84306992.mov
<img width="1791" alt="image" src="https://user-images.githubusercontent.com/3369346/203170951-d0e04645-3ab5-4ab6-858d-c6ea1a1c346b.png">
Some of the places we use `Flyout` component
```
apm • public/components/app/settings/agent_keys/create_agent_key.tsx:
99 return (
100: <EuiFlyout onClose={onCancel} size="s">
101 <EuiFlyoutHeader hasBorder>
apm • public/components/app/settings/custom_link/create_edit_custom_link_flyout/index.tsx:
78 <form onSubmit={onSubmit} id="customLink_form">
79: <EuiFlyout ownFocus onClose={onClose} size="m">
80 <EuiFlyoutHeader hasBorder>
apm • public/components/shared/apm_header_action_menu/labs/labs_flyout.tsx:
85 return (
86: <EuiFlyout onClose={onClose}>
87 <EuiFlyoutHeader hasBorder>
``` | non_priority | flyout uses full screen issue img width alt image src some of the places we use flyout component apm • public components app settings agent keys create agent key tsx return apm • public components app settings custom link create edit custom link flyout index tsx apm • public components shared apm header action menu labs labs flyout tsx return | 0 |
111,177 | 24,081,880,931 | IssuesEvent | 2022-09-19 07:26:33 | Azure/autorest.csharp | https://api.github.com/repos/Azure/autorest.csharp | reopened | DPG v2.0b2: Show Parity with DPG v1.0 | v3 Client DPG DPG/RLC v2.0b2 Epic: Parity with DPG 1.0 WS: Code Generation | One of the requirements for the [August MVP Arch Board](https://github.com/Azure/azure-sdk/issues/4549) is to show parity with DPG v1.0. This means that we can generate the same client from a Cadl file in the August MVP that we could generate from an OpenAPI file in DPG v1.0.
This should include the following generation features:
DPG **v2.0b1**:
- [x] Main namespace and assembly name **from config**
- [x] Client generation
- [x] Client constructor generation
- [x] Generation of authentication support for
- [x] KeyCredential
- [x] TokenCredential
- [x] ClientOptions generation
- [x] Subclient generation, **from .NET attributes**
- [x] Protocol methods for basic synchronous REST calls
- [x] Documentation is generated for the request/response payloads
- [x] Methods return Response
- [x] Response.Content contains serialized model
- [x] Methods support input parameters
- [x] Primitive types -> primitive types
- [x] Model types -> RequestContent
- [x] Extensible enums -> strings
- [x] Optional value types are made nullable with `?`
**August MVP Deliverable**: We should be able to generate this[ Confidential Ledger APIView](https://apiview.dev/Assemblies/Review/e25a373b908f49ab9a28436af7aede59) from a Cadl description and metadata file.
DPG **v2.0b2**:
- [ ] Paging operations
- [ ] Long-running operations
- [ ] Multiple content types
- [ ] MatchConditions: via header parameters
- [ ] Subclient generation **from Cadl**
- [ ] Main namespace and assembly name **from Cadl**
Much of this is covered by:
- https://github.com/Azure/autorest.csharp/issues/2402
- https://github.com/Azure/autorest.csharp/issues/2550 | 1.0 | DPG v2.0b2: Show Parity with DPG v1.0 - One of the requirements for the [August MVP Arch Board](https://github.com/Azure/azure-sdk/issues/4549) is to show parity with DPG v1.0. This means that we can generate the same client from a Cadl file in the August MVP that we could generate from an OpenAPI file in DPG v1.0.
This should include the following generation features:
DPG **v2.0b1**:
- [x] Main namespace and assembly name **from config**
- [x] Client generation
- [x] Client constructor generation
- [x] Generation of authentication support for
- [x] KeyCredential
- [x] TokenCredential
- [x] ClientOptions generation
- [x] Subclient generation, **from .NET attributes**
- [x] Protocol methods for basic synchronous REST calls
- [x] Documentation is generated for the request/response payloads
- [x] Methods return Response
- [x] Response.Content contains serialized model
- [x] Methods support input parameters
- [x] Primitive types -> primitive types
- [x] Model types -> RequestContent
- [x] Extensible enums -> strings
- [x] Optional value types are made nullable with `?`
**August MVP Deliverable**: We should be able to generate this[ Confidential Ledger APIView](https://apiview.dev/Assemblies/Review/e25a373b908f49ab9a28436af7aede59) from a Cadl description and metadata file.
DPG **v2.0b2**:
- [ ] Paging operations
- [ ] Long-running operations
- [ ] Multiple content types
- [ ] MatchConditions: via header parameters
- [ ] Subclient generation **from Cadl**
- [ ] Main namespace and assembly name **from Cadl**
Much of this is covered by:
- https://github.com/Azure/autorest.csharp/issues/2402
- https://github.com/Azure/autorest.csharp/issues/2550 | non_priority | dpg show parity with dpg one of the requirements for the is to show parity with dpg this means that we can generate the same client from a cadl file in the august mvp that we could generate from an openapi file in dpg this should include the following generation features dpg main namespace and assembly name from config client generation client constructor generation generation of authentication support for keycredential tokencredential clientoptions generation subclient generation from net attributes protocol methods for basic synchronous rest calls documentation is generated for the request response payloads methods return response response content contains serialized model methods support input parameters primitive types primitive types model types requestcontent extensible enums strings optional value types are made nullable with august mvp deliverable we should be able to generate this from a cadl description and metadata file dpg paging operations long running operations multiple content types matchconditions via header parameters subclient generation from cadl main namespace and assembly name from cadl much of this is covered by | 0 |
41,761 | 5,396,339,567 | IssuesEvent | 2017-02-27 11:23:20 | zkat/cacache | https://api.github.com/repos/zkat/cacache | closed | An in-range update of tap is breaking the build 🚨 | greenkeeper tests | ## Version **10.1.2** of [tap](https://github.com/tapjs/node-tap) just got published.
<table>
<tr>
<th align=left>
Branch
</th>
<td>
<a href="/zkat/cacache/compare/greenkeeper%2Ftap-10.1.2">Build failing 🚨</a>
</td>
</tr>
<tr>
<th align=left>
Dependency
</td>
<td>
tap
</td>
</tr>
<tr>
<th align=left>
Current Version
</td>
<td>
10.1.1
</td>
</tr>
<tr>
<th align=left>
Type
</td>
<td>
devDependency
</td>
</tr>
</table>
This version is **covered** by your **current version range** and after updating it in your project **the build failed**.
As tap is “only” a devDependency of this project it **might not break production or downstream projects**, but “only” your build or test tools – **preventing new deploys or publishes**.
I recommend you give this issue a high priority. I’m sure you can resolve this :muscle:
---
<details>
<summary>Status Details</summary>
- ✅ **continuous-integration/travis-ci/push** The Travis CI build passed [Details](https://travis-ci.org/zkat/cacache/builds/202839385)
- ✅ **coverage/coveralls** First build on greenkeeper/tap-10.1.2 at 94.215% [Details](https://coveralls.io/builds/10214214)
- ❌ **continuous-integration/appveyor/branch** AppVeyor build failed [Details](https://ci.appveyor.com/project/zkat/cacache/build/1.0.103)
</details>
<details>
<summary>Commits</summary>
<p>The new version differs by 3 commits .</p>
<ul>
<li><a href="https://urls.greenkeeper.io/tapjs/node-tap/commit/bf6f51ef8f4d30619ef855011ac6996295805320"><code>bf6f51e</code></a> <code>v10.1.2</code></li>
<li><a href="https://urls.greenkeeper.io/tapjs/node-tap/commit/bdb1d802bcd84d94b2da80b86bbd030ceabb9763"><code>bdb1d80</code></a> <code>Inherit bailout results from parent test</code></li>
<li><a href="https://urls.greenkeeper.io/tapjs/node-tap/commit/0aa4202ba5b04e5a5dba9c5c2e02f90b1926e3d6"><code>0aa4202</code></a> <code>Support old nodes that did not respect process.exitCode</code></li>
</ul>
<p>See the <a href="https://urls.greenkeeper.io/tapjs/node-tap/compare/9f5568d79b288f998213e63ffdc84083c8a04b77...bf6f51ef8f4d30619ef855011ac6996295805320">full diff</a>.</p>
</details>
<details>
<summary>Not sure how things should work exactly?</summary>
There is a collection of [frequently asked questions](https://greenkeeper.io/faq.html) and of course you may always [ask my humans](https://github.com/greenkeeperio/greenkeeper/issues/new).
</details>
---
Your [Greenkeeper](https://greenkeeper.io) Bot :palm_tree:
| 1.0 | An in-range update of tap is breaking the build 🚨 - ## Version **10.1.2** of [tap](https://github.com/tapjs/node-tap) just got published.
<table>
<tr>
<th align=left>
Branch
</th>
<td>
<a href="/zkat/cacache/compare/greenkeeper%2Ftap-10.1.2">Build failing 🚨</a>
</td>
</tr>
<tr>
<th align=left>
Dependency
</td>
<td>
tap
</td>
</tr>
<tr>
<th align=left>
Current Version
</td>
<td>
10.1.1
</td>
</tr>
<tr>
<th align=left>
Type
</td>
<td>
devDependency
</td>
</tr>
</table>
This version is **covered** by your **current version range** and after updating it in your project **the build failed**.
As tap is “only” a devDependency of this project it **might not break production or downstream projects**, but “only” your build or test tools – **preventing new deploys or publishes**.
I recommend you give this issue a high priority. I’m sure you can resolve this :muscle:
---
<details>
<summary>Status Details</summary>
- ✅ **continuous-integration/travis-ci/push** The Travis CI build passed [Details](https://travis-ci.org/zkat/cacache/builds/202839385)
- ✅ **coverage/coveralls** First build on greenkeeper/tap-10.1.2 at 94.215% [Details](https://coveralls.io/builds/10214214)
- ❌ **continuous-integration/appveyor/branch** AppVeyor build failed [Details](https://ci.appveyor.com/project/zkat/cacache/build/1.0.103)
</details>
<details>
<summary>Commits</summary>
<p>The new version differs by 3 commits .</p>
<ul>
<li><a href="https://urls.greenkeeper.io/tapjs/node-tap/commit/bf6f51ef8f4d30619ef855011ac6996295805320"><code>bf6f51e</code></a> <code>v10.1.2</code></li>
<li><a href="https://urls.greenkeeper.io/tapjs/node-tap/commit/bdb1d802bcd84d94b2da80b86bbd030ceabb9763"><code>bdb1d80</code></a> <code>Inherit bailout results from parent test</code></li>
<li><a href="https://urls.greenkeeper.io/tapjs/node-tap/commit/0aa4202ba5b04e5a5dba9c5c2e02f90b1926e3d6"><code>0aa4202</code></a> <code>Support old nodes that did not respect process.exitCode</code></li>
</ul>
<p>See the <a href="https://urls.greenkeeper.io/tapjs/node-tap/compare/9f5568d79b288f998213e63ffdc84083c8a04b77...bf6f51ef8f4d30619ef855011ac6996295805320">full diff</a>.</p>
</details>
<details>
<summary>Not sure how things should work exactly?</summary>
There is a collection of [frequently asked questions](https://greenkeeper.io/faq.html) and of course you may always [ask my humans](https://github.com/greenkeeperio/greenkeeper/issues/new).
</details>
---
Your [Greenkeeper](https://greenkeeper.io) Bot :palm_tree:
| non_priority | an in range update of tap is breaking the build 🚨 version of just got published branch build failing 🚨 dependency tap current version type devdependency this version is covered by your current version range and after updating it in your project the build failed as tap is “only” a devdependency of this project it might not break production or downstream projects but “only” your build or test tools – preventing new deploys or publishes i recommend you give this issue a high priority i’m sure you can resolve this muscle status details ✅ continuous integration travis ci push the travis ci build passed ✅ coverage coveralls first build on greenkeeper tap at ❌ continuous integration appveyor branch appveyor build failed commits the new version differs by commits inherit bailout results from parent test support old nodes that did not respect process exitcode see the not sure how things should work exactly there is a collection of and of course you may always your bot palm tree | 0 |
274,231 | 29,936,843,094 | IssuesEvent | 2023-06-22 13:17:31 | elastic/beats | https://api.github.com/repos/elastic/beats | opened | [Filebeat] httpjson - Allow `response.pagination` transforms to read `.last_request.*` | enhancement Filebeat Team:Security-External Integrations | **The enhancement:**
Add `.last_request.*` to the list of things that the `httpjson` input's `response.pagination` transforms can read state from.
The [existing list](https://www.elastic.co/guide/en/beats/filebeat/8.8/filebeat-input-httpjson.html#response-pagination) is: [`.last_response.*`, `.first_event.*`, `.last_event.*`, `.cursor.*`, `.header.*`, `.url.*`, `.body.*`].
**A specific use case for the enhancement:**
This would remove the need for the workaround used in the bugfix [#6649 [ti_misp] Keep the same timestamp for later pages](https://github.com/elastic/integrations/pull/6649).
The workaround for not having direct access to the last request in the pagination transformers was to add an ignored query string parameter to the actual request, since the query string parameters could be accessed via `.last_response.url.params`. | True | [Filebeat] httpjson - Allow `response.pagination` transforms to read `.last_request.*` - **The enhancement:**
Add `.last_request.*` to the list of things that the `httpjson` input's `response.pagination` transforms can read state from.
The [existing list](https://www.elastic.co/guide/en/beats/filebeat/8.8/filebeat-input-httpjson.html#response-pagination) is: [`.last_response.*`, `.first_event.*`, `.last_event.*`, `.cursor.*`, `.header.*`, `.url.*`, `.body.*`].
**A specific use case for the enhancement:**
This would remove the need for the workaround used in the bugfix [#6649 [ti_misp] Keep the same timestamp for later pages](https://github.com/elastic/integrations/pull/6649).
The workaround for not having direct access to the last request in the pagination transformers was to add an ignored query string parameter to the actual request, since the query string parameters could be accessed via `.last_response.url.params`. | non_priority | httpjson allow response pagination transforms to read last request the enhancement add last request to the list of things that the httpjson input s response pagination transforms can read state from the is a specific use case for the enhancement this would remove the need for the workaround used in the bugfix keep the same timestamp for later pages the workaround for not having direct access to the last request in the pagination transformers was to add an ignored query string parameter to the actual request since the query string parameters could be accessed via last response url params | 0 |
167,952 | 26,572,056,892 | IssuesEvent | 2023-01-21 09:34:46 | July249/shopping_mall | https://api.github.com/repos/July249/shopping_mall | closed | [issue] Price 생성 및 디자인 구현 | design Molecules | ## 👨💻 무엇을 하실 건지 설명해주세요!
- 반복적으로 사용되는 가격 컴포넌트를 molecule 단에서 생성하여 관리한다.
## 🤔 구현방법 및 예상 동작
-
## ⭐ 특이사항
-
| 1.0 | [issue] Price 생성 및 디자인 구현 - ## 👨💻 무엇을 하실 건지 설명해주세요!
- 반복적으로 사용되는 가격 컴포넌트를 molecule 단에서 생성하여 관리한다.
## 🤔 구현방법 및 예상 동작
-
## ⭐ 특이사항
-
| non_priority | price 생성 및 디자인 구현 👨💻 무엇을 하실 건지 설명해주세요 반복적으로 사용되는 가격 컴포넌트를 molecule 단에서 생성하여 관리한다 🤔 구현방법 및 예상 동작 ⭐ 특이사항 | 0 |
211,179 | 7,198,923,631 | IssuesEvent | 2018-02-05 14:28:42 | hpi-swt2/sport-portal | https://api.github.com/repos/hpi-swt2/sport-portal | reopened | Elo Ranking | epic create event epic game evaluation last sprint priority high size M team issue number 5 user story | **AS AN** organiser of a ranking list event
**I WANT TO** be able to select elo ranking as a metric
**SO THAT** the users can be ranked by this metric
## ACCEPTANCE CRITERIA
- [x] When creating a new event of the type "Ranking List" there should be a metric option "Elo"
- [x] When a user added a new match to a ranking list event the system should calculate the new elo score of the participants based on the metric of the event
- [x] The new score is calculated based on the [elo number](https://en.wikipedia.org/wiki/Elo_rating_system)
- [x] The new score of each participant is saved so that it can be displayed in the event evaluation | 1.0 | Elo Ranking - **AS AN** organiser of a ranking list event
**I WANT TO** be able to select elo ranking as a metric
**SO THAT** the users can be ranked by this metric
## ACCEPTANCE CRITERIA
- [x] When creating a new event of the type "Ranking List" there should be a metric option "Elo"
- [x] When a user added a new match to a ranking list event the system should calculate the new elo score of the participants based on the metric of the event
- [x] The new score is calculated based on the [elo number](https://en.wikipedia.org/wiki/Elo_rating_system)
- [x] The new score of each participant is saved so that it can be displayed in the event evaluation | priority | elo ranking as an organiser of a ranking list event i want to be able to select elo ranking as a metric so that the users can be ranked by this metric acceptance criteria when creating a new event of the type ranking list there should be a metric option elo when a user added a new match to a ranking list event the system should calculate the new elo score of the participants based on the metric of the event the new score is calculated based on the the new score of each participant is saved so that it can be displayed in the event evaluation | 1 |
7,283 | 24,576,082,762 | IssuesEvent | 2022-10-13 12:28:00 | elastic/apm-agent-python | https://api.github.com/repos/elastic/apm-agent-python | closed | [META 555] Add automated span type/subtype checking against shared spec | automation chore agent-python stretch | Spec PR: https://github.com/elastic/apm/pull/443
To start, we would just ensure that all span types/subtypes appear in the spec. In the future we will work on cross-agent alignment.
| 1.0 | [META 555] Add automated span type/subtype checking against shared spec - Spec PR: https://github.com/elastic/apm/pull/443
To start, we would just ensure that all span types/subtypes appear in the spec. In the future we will work on cross-agent alignment.
| non_priority | add automated span type subtype checking against shared spec spec pr to start we would just ensure that all span types subtypes appear in the spec in the future we will work on cross agent alignment | 0 |
417,756 | 28,110,942,942 | IssuesEvent | 2023-03-31 07:07:19 | KSunil2001/ped | https://api.github.com/repos/KSunil2001/ped | opened | UG is missing a few feature descriptions and incomplete command summary | severity.VeryLow type.DocumentationBug | The UG is missing the delete, sort, clear and exit feature descriptions and the command summary is missing find, help, sort and clear.
<!--session: 1680242443467-c104788c-0e4e-49b4-91c1-91a96c4dde70-->
<!--Version: Web v3.4.7--> | 1.0 | UG is missing a few feature descriptions and incomplete command summary - The UG is missing the delete, sort, clear and exit feature descriptions and the command summary is missing find, help, sort and clear.
<!--session: 1680242443467-c104788c-0e4e-49b4-91c1-91a96c4dde70-->
<!--Version: Web v3.4.7--> | non_priority | ug is missing a few feature descriptions and incomplete command summary the ug is missing the delete sort clear and exit feature descriptions and the command summary is missing find help sort and clear | 0 |
215,201 | 16,595,689,367 | IssuesEvent | 2021-06-01 13:15:22 | SOTETO/heureka | https://api.github.com/repos/SOTETO/heureka | opened | Missing base path | bug documentation | One student mentions, that s/he only sees the title of his / her new microservice, when s/he runs it as part of the Heureka!-architecture. The rest of the page stays blank. But if the microservice is called outside of the architecture, it the answer is complete.
I suggested that the base path is not configured in the application.
The base path configuration should be mentioned by the documentation! | 1.0 | Missing base path - One student mentions, that s/he only sees the title of his / her new microservice, when s/he runs it as part of the Heureka!-architecture. The rest of the page stays blank. But if the microservice is called outside of the architecture, it the answer is complete.
I suggested that the base path is not configured in the application.
The base path configuration should be mentioned by the documentation! | non_priority | missing base path one student mentions that s he only sees the title of his her new microservice when s he runs it as part of the heureka architecture the rest of the page stays blank but if the microservice is called outside of the architecture it the answer is complete
i suggested that the base path is not configured in the application the base path configuration should be mentioned by the documentation | 0 |
65,671 | 12,663,137,377 | IssuesEvent | 2020-06-18 00:19:41 | pulumi/pulumi | https://api.github.com/repos/pulumi/pulumi | closed | [Go Program Gen] Passing test for aws-eks.pp | area/codegen language/go | Will include support for namespace-less invokes, and splat expressions. | 1.0 | [Go Program Gen] Passing test for aws-eks.pp - Will include support for namespace-less invokes, and splat expressions. | non_priority | passing test for aws eks pp will include support for namespace less invokes and splat expressions | 0 |
36,884 | 12,428,127,793 | IssuesEvent | 2020-05-25 05:12:11 | mubaidr/electron-starter-template | https://api.github.com/repos/mubaidr/electron-starter-template | closed | CVE-2018-11697 High Severity Vulnerability detected by WhiteSource | security vulnerability | ## CVE-2018-11697 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.9.4</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (120)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /electron-starter-template/node_modules/node-sass/src/libsass/src/expand.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/output.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_values.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/util.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/emitter.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/lexer.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_node.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/plugins.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/base.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/position.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/operation.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /electron-starter-template/node_modules/node-sass/src/custom_importer_bridge.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/functions.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/eval.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_context_wrapper.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/node.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/emitter.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/listize.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/output.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/functions.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/cssize.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/paths.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/inspect.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/color.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/values.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/source_map.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/list.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/json.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/units.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/units.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/context.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/listize.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/string.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/context.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/boolean.h
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass2scss.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/eval.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/expand.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/factory.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/boolean.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/source_map.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/value.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /electron-starter-template/node_modules/node-sass/src/callback_bridge.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/file.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/node.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/environment.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/extend.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/operators.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/constants.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/parser.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/constants.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/list.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/cssize.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/functions.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/util.cpp
- /electron-starter-template/node_modules/node-sass/src/custom_function_bridge.cpp
- /electron-starter-template/node_modules/node-sass/src/custom_importer_bridge.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/bind.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/extend.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/debugger.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/number.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/color.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/c99func.c
- /electron-starter-template/node_modules/node-sass/src/libsass/src/position.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/values.h
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/null.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/context.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_c.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_value.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_context_wrapper.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /electron-starter-template/node_modules/node-sass/src/libsass/src/lexer.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_c.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/map.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_value.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/b64/encode.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/file.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/environment.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/plugins.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::Prelexer::exactly() which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11697>CVE-2018-11697</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-11697 High Severity Vulnerability detected by WhiteSource - ## CVE-2018-11697 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.9.4</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (120)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /electron-starter-template/node_modules/node-sass/src/libsass/src/expand.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/output.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_values.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/util.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/emitter.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/lexer.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_node.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/plugins.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/base.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/position.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/operation.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /electron-starter-template/node_modules/node-sass/src/custom_importer_bridge.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/functions.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/eval.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_context_wrapper.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/node.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/emitter.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/listize.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/output.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/functions.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/cssize.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/paths.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/inspect.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/color.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/values.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/source_map.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/list.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/json.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/units.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/units.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/context.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/listize.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/string.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/context.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/boolean.h
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass2scss.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/eval.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/expand.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/factory.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/boolean.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/source_map.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/value.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /electron-starter-template/node_modules/node-sass/src/callback_bridge.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/file.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/node.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/environment.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/extend.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/operators.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/constants.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/parser.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/constants.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/list.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/cssize.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/functions.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/util.cpp
- /electron-starter-template/node_modules/node-sass/src/custom_function_bridge.cpp
- /electron-starter-template/node_modules/node-sass/src/custom_importer_bridge.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/bind.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/extend.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/debugger.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/number.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/color.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/c99func.c
- /electron-starter-template/node_modules/node-sass/src/libsass/src/position.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/values.h
- /electron-starter-template/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/null.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/ast.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/include/sass/context.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_c.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_value.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_context_wrapper.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /electron-starter-template/node_modules/node-sass/src/libsass/src/lexer.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_c.hpp
- /electron-starter-template/node_modules/node-sass/src/sass_types/map.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/to_value.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/b64/encode.h
- /electron-starter-template/node_modules/node-sass/src/libsass/src/file.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/environment.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/plugins.hpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /electron-starter-template/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::Prelexer::exactly() which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11697>CVE-2018-11697</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries electron starter template node modules node sass src libsass src expand hpp electron starter template node modules node sass src libsass src color maps cpp electron starter template node modules node sass src libsass src sass util hpp electron starter template node modules node sass src libsass src unchecked h electron starter template node modules node sass src libsass src output hpp electron starter template node modules node sass src libsass src sass values hpp electron starter template node modules node sass src libsass src util hpp electron starter template node modules node sass src libsass src emitter hpp electron starter template node modules node sass src libsass src lexer cpp electron starter template node modules node sass src libsass test test node cpp electron starter template node modules node sass src libsass src plugins cpp electron starter template node modules node sass src libsass include sass base h electron starter template node modules node sass src libsass src position hpp electron starter template node modules node sass src libsass src subset map hpp electron starter template node modules node sass src libsass src operation hpp electron starter template node modules node sass src libsass src remove placeholders cpp electron starter template node modules node sass src libsass src error handling hpp electron starter template node modules node sass src custom importer bridge cpp electron starter template node modules node sass src libsass contrib plugin cpp electron starter template node modules node sass src libsass src functions hpp electron starter template node modules node sass src libsass test test superselector cpp electron starter template node modules node sass src libsass src eval hpp electron starter template node modules node sass src libsass src string hpp electron starter template node modules node sass src sass context wrapper h electron starter template node modules node sass src libsass src error handling cpp electron starter template node modules node sass src libsass src node cpp electron starter template node modules node sass src libsass src subset map cpp electron starter template node modules node sass src libsass src emitter cpp electron starter template node modules node sass src libsass src listize cpp electron starter template node modules node sass src libsass src ast hpp electron starter template node modules node sass src libsass src sass functions hpp electron starter template node modules node sass src libsass src memory sharedptr cpp electron starter template node modules node sass src libsass src output cpp electron starter template node modules node sass src libsass src check nesting cpp electron starter template node modules node sass src libsass src ast def macros hpp electron starter template node modules node sass src libsass src functions cpp electron starter template node modules node sass src libsass src cssize hpp electron starter template node modules node sass src libsass src prelexer cpp electron starter template node modules node sass src libsass src paths hpp electron starter template node modules node sass src libsass src ast fwd decl hpp electron starter template node modules node sass src libsass src inspect hpp electron starter template node modules node sass src sass types color cpp electron starter template node modules node sass src libsass test test unification cpp electron starter template node modules node sass src libsass src values cpp electron starter template node modules node sass src libsass src sass util cpp electron starter template node modules node sass src libsass src source map hpp electron starter template node modules node sass src sass types list h electron starter template node modules node sass src libsass src check nesting hpp electron starter template node modules node sass src libsass src json cpp electron starter template node modules node sass src libsass src units cpp electron starter template node modules node sass src libsass src units hpp electron starter template node modules node sass src libsass src context cpp electron starter template node modules node sass src libsass src checked h electron starter template node modules node sass src libsass src listize hpp electron starter template node modules node sass src sass types string cpp electron starter template node modules node sass src libsass src prelexer hpp electron starter template node modules node sass src libsass src context hpp electron starter template node modules node sass src sass types boolean h electron starter template node modules node sass src libsass include h electron starter template node modules node sass src libsass src eval cpp electron starter template node modules node sass src libsass src expand cpp electron starter template node modules node sass src sass types factory cpp electron starter template node modules node sass src sass types boolean cpp electron starter template node modules node sass src libsass src source map cpp electron starter template node modules node sass src sass types value h electron starter template node modules node sass src libsass src string cpp electron starter template node modules node sass src callback bridge h electron starter template node modules node sass src libsass src file cpp electron starter template node modules node sass src libsass src sass cpp electron starter template node modules node sass src libsass src node hpp electron starter template node modules node sass src libsass src environment cpp electron starter template node modules node sass src libsass src extend hpp electron starter template node modules node sass src libsass src sass context hpp electron starter template node modules node sass src libsass src operators hpp electron starter template node modules node sass src libsass src constants hpp electron starter template node modules node sass src libsass src sass hpp electron starter template node modules node sass src libsass src ast fwd decl cpp electron starter template node modules node sass src libsass src parser hpp electron starter template node modules node sass src libsass src constants cpp electron starter template node modules node sass src sass types list cpp electron starter template node modules node sass src libsass src cssize cpp electron starter template node modules node sass src libsass include sass functions h electron starter template node modules node sass src libsass src util cpp electron starter template node modules node sass src custom function bridge cpp electron starter template node modules node sass src custom importer bridge h electron starter template node modules node sass src libsass src bind cpp electron starter template node modules node sass src libsass src sass functions cpp electron starter template node modules node sass src libsass src backtrace cpp electron starter template node modules node sass src libsass src extend cpp electron starter template node modules node sass src sass types sass value wrapper h electron starter template node modules node sass src libsass src debugger hpp electron starter template node modules node sass src libsass src cpp electron starter template node modules node sass src sass types number cpp electron starter template node modules node sass src sass types color h electron starter template node modules node sass src libsass src c electron starter template node modules node sass src libsass src position cpp electron starter template node modules node sass src libsass src remove placeholders hpp electron starter template node modules node sass src libsass src sass values cpp electron starter template node modules node sass src libsass include sass values h electron starter template node modules node sass src libsass test test subset map cpp electron starter template node modules node sass src libsass src cpp electron starter template node modules node sass src sass types null cpp electron starter template node modules node sass src libsass src ast cpp electron starter template node modules node sass src libsass include sass context h electron starter template node modules node sass src libsass src to c cpp electron starter template node modules node sass src libsass src to value hpp electron starter template node modules node sass src libsass src color maps hpp electron starter template node modules node sass src sass context wrapper cpp electron starter template node modules node sass src libsass script test leaks pl electron starter template node modules node sass src libsass src lexer hpp electron starter template node modules node sass src libsass src memory sharedptr hpp electron starter template node modules node sass src libsass src to c hpp electron starter template node modules node sass src sass types map cpp electron starter template node modules node sass src libsass src to value cpp electron starter template node modules node sass src libsass src encode h electron starter template node modules node sass src libsass src file hpp electron starter template node modules node sass src libsass src environment hpp electron starter template node modules node sass src libsass src plugins hpp electron starter template node modules node sass src libsass src sass context cpp electron starter template node modules node sass src libsass src debug hpp vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass prelexer exactly which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
5,046 | 5,400,650,763 | IssuesEvent | 2017-02-27 22:35:32 | webpack/webpack.js.org | https://api.github.com/repos/webpack/webpack.js.org | closed | Update url's to webpack.js.org | Infrastructure | When we release webpack v2 and these new docs, we need to update `https://webpack.github.io` to `https://webpack.js.org` in many places. This issue exists to keep track of those places:
- [x] [webpack repo](https://github.com/webpack/webpack) description
- [x] webpack core code
- [x] [JS Foundation Projects page](https://js.foundation/projects/)
- [x] [webpack-dev-server readme](https://github.com/webpack/webpack-dev-server) | 1.0 | Update url's to webpack.js.org - When we release webpack v2 and these new docs, we need to update `https://webpack.github.io` to `https://webpack.js.org` in many places. This issue exists to keep track of those places:
- [x] [webpack repo](https://github.com/webpack/webpack) description
- [x] webpack core code
- [x] [JS Foundation Projects page](https://js.foundation/projects/)
- [x] [webpack-dev-server readme](https://github.com/webpack/webpack-dev-server) | non_priority | update url s to webpack js org when we release webpack and these new docs we need to update to in many places this issue exists to keep track of those places description webpack core code | 0 |
106,544 | 4,274,234,484 | IssuesEvent | 2016-07-13 19:51:56 | docker/docker | https://api.github.com/repos/docker/docker | closed | [1.12] evaluate new single-letter flags before GA | area/cli priority/P1 | With the SwarmKit integration, many new commands and options were added; we should be careful when adding shorthand (single-letter) flags to commands, and only use shorthand flags for frequently used options.
In doubt; remove the shorthand flag, so that we can decide later when really needed.
Opening this issue so that we don't forget to evaluate all new flags before GA | 1.0 | [1.12] evaluate new single-letter flags before GA - With the SwarmKit integration, many new commands and options were added; we should be careful when adding shorthand (single-letter) flags to commands, and only use shorthand flags for frequently used options.
In doubt; remove the shorthand flag, so that we can decide later when really needed.
Opening this issue so that we don't forget to evaluate all new flags before GA | priority | evaluate new single letter flags before ga with the swarmkit integration many new commands and options were added we should be careful when adding shorthand single letter flags to commands and only use shorthand flags for frequently used options in doubt remove the shorthand flag so that we can decide later when really needed opening this issue so that we don t forget to evaluate all new flags before ga | 1 |
56,326 | 8,066,959,820 | IssuesEvent | 2018-08-04 22:51:33 | ProjectEvergreen/project-evergreen | https://api.github.com/repos/ProjectEvergreen/project-evergreen | closed | Add support for unit testing | documentation enhancement help wanted question todo-app website | - [**Polymer CLI**](https://github.com/Polymer/tools/tree/master/packages/cli#features) uses [**web-component-tester**](https://github.com/Polymer/tools/tree/master/packages/web-component-tester) (which uses **mocha**) - not sure if this is compatible with "vanilla" web compoennts (not using Polymer / Polymer CLI) | 1.0 | Add support for unit testing - - [**Polymer CLI**](https://github.com/Polymer/tools/tree/master/packages/cli#features) uses [**web-component-tester**](https://github.com/Polymer/tools/tree/master/packages/web-component-tester) (which uses **mocha**) - not sure if this is compatible with "vanilla" web compoennts (not using Polymer / Polymer CLI) | non_priority | add support for unit testing uses which uses mocha not sure if this is compatible with vanilla web compoennts not using polymer polymer cli | 0 |
71,600 | 9,530,267,156 | IssuesEvent | 2019-04-29 13:31:01 | roboticslab-uc3m/teo-developer-manual | https://api.github.com/repos/roboticslab-uc3m/teo-developer-manual | closed | Export TEO Network Information to labs repo | documentation | Export TEO Network Information to labs repo: https://github.com/roboticslab-uc3m/teo-developer-manual/blob/a727cda044a9796951f0a63aa72ce84325b65a6c/appendix/c-teo-network-information.md
Essentially only manipulation, locomotion and head make sense here. | 1.0 | Export TEO Network Information to labs repo - Export TEO Network Information to labs repo: https://github.com/roboticslab-uc3m/teo-developer-manual/blob/a727cda044a9796951f0a63aa72ce84325b65a6c/appendix/c-teo-network-information.md
Essentially only manipulation, locomotion and head make sense here. | non_priority | export teo network information to labs repo export teo network information to labs repo essentially only manipulation locomotion and head make sense here | 0 |
716,159 | 24,624,060,605 | IssuesEvent | 2022-10-16 09:27:47 | joomlahenk/fabrik | https://api.github.com/repos/joomlahenk/fabrik | closed | Fatal error trying to add a new element | High Priority | Just trying to add a new element. This happens since latest GH update:
Failed opening required '/var/www/html/j4test/plugins/fabrik_element/radiobutton/radiobutton.php' (include_path='.:/usr/share/php')
I have no radiobutton element in the list. | 1.0 | Fatal error trying to add a new element - Just trying to add a new element. This happens since latest GH update:
Failed opening required '/var/www/html/j4test/plugins/fabrik_element/radiobutton/radiobutton.php' (include_path='.:/usr/share/php')
I have no radiobutton element in the list. | priority | fatal error trying to add a new element just trying to add a new element this happens since latest gh update failed opening required var www html plugins fabrik element radiobutton radiobutton php include path usr share php i have no radiobutton element in the list | 1 |
471,281 | 13,563,920,169 | IssuesEvent | 2020-09-18 09:16:15 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | StyleEditor enableSetDefaultStyle property does not work | Accepted Priority: High Project: C183 StyleEditor bug | ## Description
<!-- Add here a few sentences describing the bug. -->
StyleEditor should display a button to set a style as default.
It was possible to activate this functionality with the property `enableSetDefaultStyle` set to true:
```js
{
"name": "StyleEditor",
"cfg": {
"enableSetDefaultStyle": true
}
}
```
From a quick investigation it seems that the StyleToolbar inside the TOCItemsSettings tabs does not get the localConfig cfg object.
https://github.com/geosolutions-it/MapStore2/blob/fbb6d99f99e0114d43725046f0379894ba7c026f/web/client/plugins/tocitemssettings/defaultSettingsTabs.js#L194
## How to reproduce
<!-- A list of steps to reproduce the bug -->
- add enableSetDefaultStyle property to localConfig.json
- login as admin
- open the map viewer and add a vector layer from gs-stable
- select the layer in TOC and opent the settings
- select the style tab
*Expected Result*
<!-- Describe here the expected result -->
A button with a star icon should be available in the top toolbar
*Current Result*
<!-- Describe here the current behavior -->
The button to set a style as default is not available
- [x] Not browser related
<details><summary> <b>Browser info</b> </summary>
<!-- If browser related, please compile the following table -->
<!-- If your browser is not in the list please add a new row to the table with the version -->
(use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users)
| Browser Affected | Version |
|---|---|
|Internet Explorer| |
|Edge| |
|Chrome| |
|Firefox| |
|Safari| |
</details>
## Other useful information
<!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
| 1.0 | StyleEditor enableSetDefaultStyle property does not work - ## Description
<!-- Add here a few sentences describing the bug. -->
StyleEditor should display a button to set a style as default.
It was possible to activate this functionality with the property `enableSetDefaultStyle` set to true:
```js
{
"name": "StyleEditor",
"cfg": {
"enableSetDefaultStyle": true
}
}
```
From a quick investigation it seems that the StyleToolbar inside the TOCItemsSettings tabs does not get the localConfig cfg object.
https://github.com/geosolutions-it/MapStore2/blob/fbb6d99f99e0114d43725046f0379894ba7c026f/web/client/plugins/tocitemssettings/defaultSettingsTabs.js#L194
## How to reproduce
<!-- A list of steps to reproduce the bug -->
- add enableSetDefaultStyle property to localConfig.json
- login as admin
- open the map viewer and add a vector layer from gs-stable
- select the layer in TOC and opent the settings
- select the style tab
*Expected Result*
<!-- Describe here the expected result -->
A button with a star icon should be available in the top toolbar
*Current Result*
<!-- Describe here the current behavior -->
The button to set a style as default is not available
- [x] Not browser related
<details><summary> <b>Browser info</b> </summary>
<!-- If browser related, please compile the following table -->
<!-- If your browser is not in the list please add a new row to the table with the version -->
(use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users)
| Browser Affected | Version |
|---|---|
|Internet Explorer| |
|Edge| |
|Chrome| |
|Firefox| |
|Safari| |
</details>
## Other useful information
<!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
| priority | styleeditor enablesetdefaultstyle property does not work description styleeditor should display a button to set a style as default it was possible to activate this functionality with the property enablesetdefaultstyle set to true js name styleeditor cfg enablesetdefaultstyle true from a quick investigation it seems that the styletoolbar inside the tocitemssettings tabs does not get the localconfig cfg object how to reproduce add enablesetdefaultstyle property to localconfig json login as admin open the map viewer and add a vector layer from gs stable select the layer in toc and opent the settings select the style tab expected result a button with a star icon should be available in the top toolbar current result the button to set a style as default is not available not browser related browser info use this site a href for non expert users browser affected version internet explorer edge chrome firefox safari other useful information | 1 |
736,118 | 25,459,145,791 | IssuesEvent | 2022-11-24 16:45:54 | googleapis/doc-pipeline | https://api.github.com/repos/googleapis/doc-pipeline | opened | Flake8 test failing | type: bug priority: p2 | Error stack:
```
Traceback (most recent call last):
File "/h/.local/bin/flake8", line 8, in <module>
sys.exit(main())
^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/main/cli.py", line 23, in main
app.run(argv)
File "/h/.local/lib/python3.11/site-packages/flake8/main/application.py", line 198, in run
self._run(argv)
File "/h/.local/lib/python3.11/site-packages/flake8/main/application.py", line 186, in _run
self.initialize(argv)
File "/h/.local/lib/python3.11/site-packages/flake8/main/application.py", line 165, in initialize
self.plugins, self.options = parse_args(argv)
^^^^^^^^^^^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/options/parse_args.py", line 53, in parse_args
opts = aggregator.aggregate_options(option_manager, cfg, cfg_dir, rest)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/options/aggregator.py", line 30, in aggregate_options
parsed_config = config.parse_config(manager, cfg, cfg_dir)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/options/config.py", line 131, in parse_config
raise ValueError(
ValueError: Error code '#' supplied to 'ignore' option does not match '^[A-Z]{1,3}[0-9]{0,3}$'
```
according to [Stackoverflow post](https://stackoverflow.com/questions/74558565/flake8-error-code-supplied-to-ignore-option-does-not-match-a-z1-30), we can't have comments in the same line, must be in a separate line since `flake>=6`. | 1.0 | Flake8 test failing - Error stack:
```
Traceback (most recent call last):
File "/h/.local/bin/flake8", line 8, in <module>
sys.exit(main())
^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/main/cli.py", line 23, in main
app.run(argv)
File "/h/.local/lib/python3.11/site-packages/flake8/main/application.py", line 198, in run
self._run(argv)
File "/h/.local/lib/python3.11/site-packages/flake8/main/application.py", line 186, in _run
self.initialize(argv)
File "/h/.local/lib/python3.11/site-packages/flake8/main/application.py", line 165, in initialize
self.plugins, self.options = parse_args(argv)
^^^^^^^^^^^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/options/parse_args.py", line 53, in parse_args
opts = aggregator.aggregate_options(option_manager, cfg, cfg_dir, rest)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/options/aggregator.py", line 30, in aggregate_options
parsed_config = config.parse_config(manager, cfg, cfg_dir)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/h/.local/lib/python3.11/site-packages/flake8/options/config.py", line 131, in parse_config
raise ValueError(
ValueError: Error code '#' supplied to 'ignore' option does not match '^[A-Z]{1,3}[0-9]{0,3}$'
```
according to [Stackoverflow post](https://stackoverflow.com/questions/74558565/flake8-error-code-supplied-to-ignore-option-does-not-match-a-z1-30), we can't have comments in the same line, must be in a separate line since `flake>=6`. | priority | test failing error stack traceback most recent call last file h local bin line in sys exit main file h local lib site packages main cli py line in main app run argv file h local lib site packages main application py line in run self run argv file h local lib site packages main application py line in run self initialize argv file h local lib site packages main application py line in initialize self plugins self options parse args argv file h local lib site packages options parse args py line in parse args opts aggregator aggregate options option manager cfg cfg dir rest file h local lib site packages options aggregator py line in aggregate options parsed config config parse config manager cfg cfg dir file h local lib site packages options config py line in parse config raise valueerror valueerror error code supplied to ignore option does not match according to we can t have comments in the same line must be in a separate line since flake | 1 |
99,003 | 11,102,726,064 | IssuesEvent | 2019-12-17 01:07:37 | Dev-Eritas/Dev-Eritas | https://api.github.com/repos/Dev-Eritas/Dev-Eritas | opened | Investigar sobre integración continua con Github Actions | analisis back-end documentation sprint5 tercera parte | ## User Story
**COMO** Dev-Eritas
**QUIERO** Saber como funcionan las acciones de Github
**PARA QUE** Se pueda aplicar la integración continua en el proyecto
## Criterio de Aceptacion
- [ ] Conocer como aplicar github actions
| 1.0 | Investigar sobre integración continua con Github Actions - ## User Story
**COMO** Dev-Eritas
**QUIERO** Saber como funcionan las acciones de Github
**PARA QUE** Se pueda aplicar la integración continua en el proyecto
## Criterio de Aceptacion
- [ ] Conocer como aplicar github actions
| non_priority | investigar sobre integración continua con github actions user story como dev eritas quiero saber como funcionan las acciones de github para que se pueda aplicar la integración continua en el proyecto criterio de aceptacion conocer como aplicar github actions | 0 |
65,236 | 19,279,115,817 | IssuesEvent | 2021-12-10 15:14:14 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Sign out leads to "Something went wrong!" page. | T-Defect | ### Steps to reproduce
Be logged in to a server previously in that browser (matrix.org and a test server have both exhibited this).
Visit app.element.io ; get to the home screen for the previously logged in user.
Head into the menu and click log out. Don't otherwise interact with the page.
### Outcome
#### What did you expect?
Not to be asked to report a bug. maybe a "logged out" message, or the intial page again.
#### What happened instead?
"Something went wrong!" message.
Note on firefox I appear to be unable to send debug logging. On firefox this occurred when using a private browsing session.
### Operating system
Linux ubuntu
### Browser information
Firefox and Chrome
### URL for webapp
app.element.io
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
Yes | 1.0 | Sign out leads to "Something went wrong!" page. - ### Steps to reproduce
Be logged in to a server previously in that browser (matrix.org and a test server have both exhibited this).
Visit app.element.io ; get to the home screen for the previously logged in user.
Head into the menu and click log out. Don't otherwise interact with the page.
### Outcome
#### What did you expect?
Not to be asked to report a bug. maybe a "logged out" message, or the intial page again.
#### What happened instead?
"Something went wrong!" message.
Note on firefox I appear to be unable to send debug logging. On firefox this occurred when using a private browsing session.
### Operating system
Linux ubuntu
### Browser information
Firefox and Chrome
### URL for webapp
app.element.io
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
Yes | non_priority | sign out leads to something went wrong page steps to reproduce be logged in to a server previously in that browser matrix org and a test server have both exhibited this visit app element io get to the home screen for the previously logged in user head into the menu and click log out don t otherwise interact with the page outcome what did you expect not to be asked to report a bug maybe a logged out message or the intial page again what happened instead something went wrong message note on firefox i appear to be unable to send debug logging on firefox this occurred when using a private browsing session operating system linux ubuntu browser information firefox and chrome url for webapp app element io application version no response homeserver no response will you send logs yes | 0 |
661,331 | 22,049,310,190 | IssuesEvent | 2022-05-30 07:07:46 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | xhamster.com - video or audio doesn't play | priority-important browser-focus-geckoview engine-gecko | <!-- @browser: Firefox Mobile 100.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:100.0) Gecko/100.0 Firefox/100.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/105099 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://xhamster.com/gay/categories/american
**Browser / Version**: Firefox Mobile 100.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Other
**Problem type**: Video or audio doesn't play
**Description**: There is no video
**Steps to Reproduce**:
No video. Only audio
Pls check on this
<details>
<summary>View the screenshot</summary>
**Screenshot removed - explicit content**
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220519220738</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/5/c29b5bc6-8c16-4a1c-9dd2-93043ef3accd)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | xhamster.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 100.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:100.0) Gecko/100.0 Firefox/100.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/105099 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://xhamster.com/gay/categories/american
**Browser / Version**: Firefox Mobile 100.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Other
**Problem type**: Video or audio doesn't play
**Description**: There is no video
**Steps to Reproduce**:
No video. Only audio
Pls check on this
<details>
<summary>View the screenshot</summary>
**Screenshot removed - explicit content**
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220519220738</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/5/c29b5bc6-8c16-4a1c-9dd2-93043ef3accd)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | xhamster com video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes other problem type video or audio doesn t play description there is no video steps to reproduce no video only audio pls check on this view the screenshot screenshot removed explicit content browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
138,456 | 5,341,475,765 | IssuesEvent | 2017-02-17 03:03:45 | mmisw/mmiorr | https://api.github.com/repos/mmisw/mmiorr | closed | search mappings by given confidence level | enhancement imported Priority-Low vine wontfix | _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on November 19, 2008 08:22:36_
TBD What capability do you want added or improved? Where do you want this capability to be accessible? What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies,
text contents of vocabulary (voc2rdf), operating system, browser/version
(Firefox, Safari, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=68_
| 1.0 | search mappings by given confidence level - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on November 19, 2008 08:22:36_
TBD What capability do you want added or improved? Where do you want this capability to be accessible? What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies,
text contents of vocabulary (voc2rdf), operating system, browser/version
(Firefox, Safari, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=68_
| priority | search mappings by given confidence level from on november tbd what capability do you want added or improved where do you want this capability to be accessible what sort of input command mechanism do you want what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari ie etc screenshot etc original issue | 1 |
430,662 | 12,464,506,033 | IssuesEvent | 2020-05-28 12:36:01 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Sync problems after upgrading to 0.32.5 | Administration/Metadata & Sync Priority:P3 Type:Bug | Recently upgraded to 0.32.5, and it appears that the DB migration did not go exactly as it should.
After migration I could not get sync to work properly. E.g. in one particular table with lots of columns, only a small subset would show up under data model. Any number of explicit sync requests would not make any difference.
Looking more closely in the internal metabase database, I noticed that the missing columns appeared to have a weird state in the metabase_field table: all the missing columns had active = 'true' and visibility_type='retired'.
Explicitly changing visibility_type to 'normal' for these fields fixed the problem (for this particular table).
I'm just guessing here, but maybe something is missing in a DB migration executed on upgrade to 0.32.5? | 1.0 | Sync problems after upgrading to 0.32.5 - Recently upgraded to 0.32.5, and it appears that the DB migration did not go exactly as it should.
After migration I could not get sync to work properly. E.g. in one particular table with lots of columns, only a small subset would show up under data model. Any number of explicit sync requests would not make any difference.
Looking more closely in the internal metabase database, I noticed that the missing columns appeared to have a weird state in the metabase_field table: all the missing columns had active = 'true' and visibility_type='retired'.
Explicitly changing visibility_type to 'normal' for these fields fixed the problem (for this particular table).
I'm just guessing here, but maybe something is missing in a DB migration executed on upgrade to 0.32.5? | priority | sync problems after upgrading to recently upgraded to and it appears that the db migration did not go exactly as it should after migration i could not get sync to work properly e g in one particular table with lots of columns only a small subset would show up under data model any number of explicit sync requests would not make any difference looking more closely in the internal metabase database i noticed that the missing columns appeared to have a weird state in the metabase field table all the missing columns had active true and visibility type retired explicitly changing visibility type to normal for these fields fixed the problem for this particular table i m just guessing here but maybe something is missing in a db migration executed on upgrade to | 1 |
703,996 | 24,180,535,967 | IssuesEvent | 2022-09-23 08:28:16 | authelia/authelia | https://api.github.com/repos/authelia/authelia | opened | 2FA, TOTP, Failed to register device | priority/4/normal type/bug/unconfirmed status/needs-triage | ### Version
v4.36.8
### Deployment Method
Docker
### Reverse Proxy
Caddy
### Reverse Proxy Version
2.6.1
### Description
When I try to register a device for TOTP 2FA the link is always expired, even if I open it seconds after the request:


1FA works fine. I also verified that Timezone and date are correct.
### Reproduction
1. Go to https://2fa.test.<redacted>.de
2. Get redirected to the Authelia login page
3. Login using username and password
4. Get the register device screen
5. Click the register device link
6. Open the link from notifications.txt
7. See the error shown in the screenshots
### Expectations
Being able to register a device for 2FA via TOTP
### Logs
```shell
authelia | time="2022-09-23T09:52:15+02:00" level=debug msg="Check authorization of subject username= groups= ip=123.456.789.123 and object https://2fa.test.example.com/ (method GET)."
authelia | time="2022-09-23T09:52:15+02:00" level=info msg="Access to https://2fa.test.example.com/ (method GET) is not authorized to user <anonymous>, responding with status code 302 with location redirect to https://auth.test.example.com/?rd=https%3A%2F%2F2fa.test.example.com%2F&rm=GET" method=GET path=/api/verify remote_ip=
123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Mark 1FA authentication attempt made by user '<redacted>'" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Successful 1FA authentication attempt made by user '<redacted>'" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Check authorization of subject username=<redacted> groups=mygroup ip=123.456.789.123 and object https://2fa.test.example.com/ (method GET)."
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Required level for the URL https://2fa.test.example.com/ is 2" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=warning msg="https://2fa.test.example.com/ requires 2FA, cannot be redirected yet" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:24+02:00" level=debug msg="Sending an email to user user1 (example@example.com) to confirm identity for registering a device." method=POST path=/api/secondfactor/totp/identity/start remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:39+02:00" level=error msg="Cannot parse token" method=POST path=/api/secondfactor/totp/identity/finish remote_ip=123.456.789.123 stack="github.com/authelia/authelia/v4/internal/middlewares/authelia_context.go:66 (*AutheliaCtx).Error\ngithub.com/authelia/authelia/v4/internal/middle
wares/identity_verification.go:143 IdentityVerificationFinish.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/require_first_factor.go:15 Require1FA.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/bridge.go:54 (*BridgeBuilder).Build.func1.1\ngithub.com/authelia/authelia/v4/inter
nal/middlewares/headers.go:25 SecurityHeadersCSPNone.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/headers.go:35 SecurityHeadersNoStore.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/headers.go:16 SecurityHeaders.func1\ngithub.com/fasthttp/router@v
1.4.12/router.go:414 (*Router).Handler\ngithub.com/valyala/fasthttp@v1.40.0/http.go:154 (*Response).StatusCode\ngithub.com/valyala/fasthttp@v1.40.0/server.go:2311 (*Server).serveConn\ngithub.com/valyala/fasthttp@v1.40.0/wo
rkerpool.go:224 (*workerPool).workerFunc\ngithub.com/valyala/fasthttp@v1.40.0/workerpool.go:196 (*workerPool).getCh.func1\nruntime/asm_amd64.s:1594 goexit"
```
### Configuration
```yaml
---
###############################################################
# Authelia configuration #
###############################################################
server:
host: 0.0.0.0
port: 9091
log:
level: debug
theme: dark
#jwt_secret: a_very_important_secret ---> set via ENV VAR AUTHELIA_JWT_SECRET
default_redirection_url: https://auth.test.example.com
totp:
issuer: authelia.com
authentication_backend:
# Password reset through authelia works normally.
password_reset:
disable: false
# How often authelia should check if there is an user update in LDAP
refresh_interval: 1m
ldap:
implementation: custom
# Pattern is ldap://HOSTNAME-OR-IP:PORT
# Normal ldap port is 389, standard in LLDAP is 3890
url: ldap://lldap:3890
# The dial timeout for LDAP.
timeout: 5s
# Use StartTLS with the LDAP connection, TLS not supported right now
start_tls: false
#tls:
# skip_verify: false
# minimum_version: TLS1.2
# Set base dn, like dc=google,dc.com
# base_dn: dc=example,dc=com ---> Set via ENV VAR AUTHELIA_AUTHENTICATION_BACKEND_LDAP_BASE_DN
username_attribute: uid
# You need to set this to ou=people, because all users are stored in this ou!
additional_users_dn: ou=people
# To allow sign in both with username and email, one can use a filter like
# (&(|({username_attribute}={input})({mail_attribute}={input}))(objectClass=person))
users_filter: (&({username_attribute}={input})(objectClass=person))
# Set this to ou=groups, because all groups are stored in this ou
additional_groups_dn: ou=groups
# Only this filter is supported right now
groups_filter: (member={dn})
# The attribute holding the name of the group.
group_name_attribute: cn
# Email attribute
mail_attribute: mail
# The attribute holding the display name of the user. This will be used to greet an authenticated user.
display_name_attribute: displayName
# The username and password of the admin user.
# "admin" should be the admin username you set in the LLDAP configuration
#user: uid=admin,ou=people,dc=example,dc=com ---> Set via ENV VAR AUTHELIA_AUTHENTICATION_BACKEND_LDAP_USER
#password: "REPLACE_ME" ---> Set via ENV VAR AUTHELIA_AUTHENTICATION_BACKEND_LDAP_PASSWORD
access_control:
default_policy: deny
rules:
# Rules applied to everyone
- domain: public.test.example.com
policy: bypass
- domain: 1fa.test.example.com
policy: one_factor
- domain: 2fa.test.example.com
policy: two_factor
session:
name: authelia_session
# secret: unsecure_session_secret ---> Set via ENV VAR AUTHELIA_SESSION_SECRET
expiration: 3600 # 1 hour
inactivity: 300 # 5 minutes
domain: example.com # Should match whatever your root protected domain is
regulation:
max_retries: 3
find_time: 120
ban_time: 300
storage:
local:
path: /config/db.sqlite3
notifier:
filesystem:
filename: /config/notification.txt
```
### Documentation
Here's the Caddyfile
```Caddyfile
{
debug
log {
format console
}
}
(trusted_proxy_list) {
## Uncomment & adjust the following line to configure specific ranges which should be considered as trustworthy.
trusted_proxies 10.0.0.0/8 172.16.0.0/16 192.168.0.0/16 fc00::/7
}
(authelia) {
forward_auth authelia:9091 {
uri /api/verify?rd=https://auth.test.example.com/
copy_headers Remote-User Remote-Groups Remote-Name Remote-Email
## This import needs to be included if you're relying on a trusted proxies configuration.
import trusted_proxy_list
}
}
auth.test.example.com {
reverse_proxy authelia:9091 {
## This import needs to be included if you're relying on a trusted proxies configuration.
import trusted_proxy_list
}
}
lldap.test.example.com {
reverse_proxy lldap:17170
}
public.test.example.com {
import authelia
respond "Public"
}
1fa.test.example.com {
import authelia
respond "1FA"
}
2fa.test.example.com {
import authelia
respond "2FA"
}
``` | 1.0 | 2FA, TOTP, Failed to register device - ### Version
v4.36.8
### Deployment Method
Docker
### Reverse Proxy
Caddy
### Reverse Proxy Version
2.6.1
### Description
When I try to register a device for TOTP 2FA the link is always expired, even if I open it seconds after the request:


1FA works fine. I also verified that Timezone and date are correct.
### Reproduction
1. Go to https://2fa.test.<redacted>.de
2. Get redirected to the Authelia login page
3. Login using username and password
4. Get the register device screen
5. Click the register device link
6. Open the link from notifications.txt
7. See the error shown in the screenshots
### Expectations
Being able to register a device for 2FA via TOTP
### Logs
```shell
authelia | time="2022-09-23T09:52:15+02:00" level=debug msg="Check authorization of subject username= groups= ip=123.456.789.123 and object https://2fa.test.example.com/ (method GET)."
authelia | time="2022-09-23T09:52:15+02:00" level=info msg="Access to https://2fa.test.example.com/ (method GET) is not authorized to user <anonymous>, responding with status code 302 with location redirect to https://auth.test.example.com/?rd=https%3A%2F%2F2fa.test.example.com%2F&rm=GET" method=GET path=/api/verify remote_ip=
123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Mark 1FA authentication attempt made by user '<redacted>'" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Successful 1FA authentication attempt made by user '<redacted>'" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Check authorization of subject username=<redacted> groups=mygroup ip=123.456.789.123 and object https://2fa.test.example.com/ (method GET)."
authelia | time="2022-09-23T09:52:21+02:00" level=debug msg="Required level for the URL https://2fa.test.example.com/ is 2" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:21+02:00" level=warning msg="https://2fa.test.example.com/ requires 2FA, cannot be redirected yet" method=POST path=/api/firstfactor remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:24+02:00" level=debug msg="Sending an email to user user1 (example@example.com) to confirm identity for registering a device." method=POST path=/api/secondfactor/totp/identity/start remote_ip=123.456.789.123
authelia | time="2022-09-23T09:52:39+02:00" level=error msg="Cannot parse token" method=POST path=/api/secondfactor/totp/identity/finish remote_ip=123.456.789.123 stack="github.com/authelia/authelia/v4/internal/middlewares/authelia_context.go:66 (*AutheliaCtx).Error\ngithub.com/authelia/authelia/v4/internal/middle
wares/identity_verification.go:143 IdentityVerificationFinish.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/require_first_factor.go:15 Require1FA.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/bridge.go:54 (*BridgeBuilder).Build.func1.1\ngithub.com/authelia/authelia/v4/inter
nal/middlewares/headers.go:25 SecurityHeadersCSPNone.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/headers.go:35 SecurityHeadersNoStore.func1\ngithub.com/authelia/authelia/v4/internal/middlewares/headers.go:16 SecurityHeaders.func1\ngithub.com/fasthttp/router@v
1.4.12/router.go:414 (*Router).Handler\ngithub.com/valyala/fasthttp@v1.40.0/http.go:154 (*Response).StatusCode\ngithub.com/valyala/fasthttp@v1.40.0/server.go:2311 (*Server).serveConn\ngithub.com/valyala/fasthttp@v1.40.0/wo
rkerpool.go:224 (*workerPool).workerFunc\ngithub.com/valyala/fasthttp@v1.40.0/workerpool.go:196 (*workerPool).getCh.func1\nruntime/asm_amd64.s:1594 goexit"
```
### Configuration
```yaml
---
###############################################################
# Authelia configuration #
###############################################################
server:
host: 0.0.0.0
port: 9091
log:
level: debug
theme: dark
#jwt_secret: a_very_important_secret ---> set via ENV VAR AUTHELIA_JWT_SECRET
default_redirection_url: https://auth.test.example.com
totp:
issuer: authelia.com
authentication_backend:
# Password reset through authelia works normally.
password_reset:
disable: false
# How often authelia should check if there is an user update in LDAP
refresh_interval: 1m
ldap:
implementation: custom
# Pattern is ldap://HOSTNAME-OR-IP:PORT
# Normal ldap port is 389, standard in LLDAP is 3890
url: ldap://lldap:3890
# The dial timeout for LDAP.
timeout: 5s
# Use StartTLS with the LDAP connection, TLS not supported right now
start_tls: false
#tls:
# skip_verify: false
# minimum_version: TLS1.2
# Set base dn, like dc=google,dc.com
# base_dn: dc=example,dc=com ---> Set via ENV VAR AUTHELIA_AUTHENTICATION_BACKEND_LDAP_BASE_DN
username_attribute: uid
# You need to set this to ou=people, because all users are stored in this ou!
additional_users_dn: ou=people
# To allow sign in both with username and email, one can use a filter like
# (&(|({username_attribute}={input})({mail_attribute}={input}))(objectClass=person))
users_filter: (&({username_attribute}={input})(objectClass=person))
# Set this to ou=groups, because all groups are stored in this ou
additional_groups_dn: ou=groups
# Only this filter is supported right now
groups_filter: (member={dn})
# The attribute holding the name of the group.
group_name_attribute: cn
# Email attribute
mail_attribute: mail
# The attribute holding the display name of the user. This will be used to greet an authenticated user.
display_name_attribute: displayName
# The username and password of the admin user.
# "admin" should be the admin username you set in the LLDAP configuration
#user: uid=admin,ou=people,dc=example,dc=com ---> Set via ENV VAR AUTHELIA_AUTHENTICATION_BACKEND_LDAP_USER
#password: "REPLACE_ME" ---> Set via ENV VAR AUTHELIA_AUTHENTICATION_BACKEND_LDAP_PASSWORD
access_control:
default_policy: deny
rules:
# Rules applied to everyone
- domain: public.test.example.com
policy: bypass
- domain: 1fa.test.example.com
policy: one_factor
- domain: 2fa.test.example.com
policy: two_factor
session:
name: authelia_session
# secret: unsecure_session_secret ---> Set via ENV VAR AUTHELIA_SESSION_SECRET
expiration: 3600 # 1 hour
inactivity: 300 # 5 minutes
domain: example.com # Should match whatever your root protected domain is
regulation:
max_retries: 3
find_time: 120
ban_time: 300
storage:
local:
path: /config/db.sqlite3
notifier:
filesystem:
filename: /config/notification.txt
```
### Documentation
Here's the Caddyfile
```Caddyfile
{
debug
log {
format console
}
}
(trusted_proxy_list) {
## Uncomment & adjust the following line to configure specific ranges which should be considered as trustworthy.
trusted_proxies 10.0.0.0/8 172.16.0.0/16 192.168.0.0/16 fc00::/7
}
(authelia) {
forward_auth authelia:9091 {
uri /api/verify?rd=https://auth.test.example.com/
copy_headers Remote-User Remote-Groups Remote-Name Remote-Email
## This import needs to be included if you're relying on a trusted proxies configuration.
import trusted_proxy_list
}
}
auth.test.example.com {
reverse_proxy authelia:9091 {
## This import needs to be included if you're relying on a trusted proxies configuration.
import trusted_proxy_list
}
}
lldap.test.example.com {
reverse_proxy lldap:17170
}
public.test.example.com {
import authelia
respond "Public"
}
1fa.test.example.com {
import authelia
respond "1FA"
}
2fa.test.example.com {
import authelia
respond "2FA"
}
``` | priority | totp failed to register device version deployment method docker reverse proxy caddy reverse proxy version description when i try to register a device for totp the link is always expired even if i open it seconds after the request works fine i also verified that timezone and date are correct reproduction go to get redirected to the authelia login page login using username and password get the register device screen click the register device link open the link from notifications txt see the error shown in the screenshots expectations being able to register a device for via totp logs shell authelia time level debug msg check authorization of subject username groups ip and object method get authelia time level info msg access to method get is not authorized to user responding with status code with location redirect to method get path api verify remote ip authelia time level debug msg mark authentication attempt made by user method post path api firstfactor remote ip authelia time level debug msg successful authentication attempt made by user method post path api firstfactor remote ip authelia time level debug msg check authorization of subject username groups mygroup ip and object method get authelia time level debug msg required level for the url is method post path api firstfactor remote ip authelia time level warning msg requires cannot be redirected yet method post path api firstfactor remote ip authelia time level debug msg sending an email to user example example com to confirm identity for registering a device method post path api secondfactor totp identity start remote ip authelia time level error msg cannot parse token method post path api secondfactor totp identity finish remote ip stack github com authelia authelia internal middlewares authelia context go autheliactx error ngithub com authelia authelia internal middle wares identity verification go identityverificationfinish ngithub com authelia authelia internal middlewares require first factor go ngithub com authelia authelia internal middlewares bridge go bridgebuilder build ngithub com authelia authelia inter nal middlewares headers go securityheaderscspnone ngithub com authelia authelia internal middlewares headers go securityheadersnostore ngithub com authelia authelia internal middlewares headers go securityheaders ngithub com fasthttp router v router go router handler ngithub com valyala fasthttp http go response statuscode ngithub com valyala fasthttp server go server serveconn ngithub com valyala fasthttp wo rkerpool go workerpool workerfunc ngithub com valyala fasthttp workerpool go workerpool getch nruntime asm s goexit configuration yaml authelia configuration server host port log level debug theme dark jwt secret a very important secret set via env var authelia jwt secret default redirection url totp issuer authelia com authentication backend password reset through authelia works normally password reset disable false how often authelia should check if there is an user update in ldap refresh interval ldap implementation custom pattern is ldap hostname or ip port normal ldap port is standard in lldap is url ldap lldap the dial timeout for ldap timeout use starttls with the ldap connection tls not supported right now start tls false tls skip verify false minimum version set base dn like dc google dc com base dn dc example dc com set via env var authelia authentication backend ldap base dn username attribute uid you need to set this to ou people because all users are stored in this ou additional users dn ou people to allow sign in both with username and email one can use a filter like username attribute input mail attribute input objectclass person users filter username attribute input objectclass person set this to ou groups because all groups are stored in this ou additional groups dn ou groups only this filter is supported right now groups filter member dn the attribute holding the name of the group group name attribute cn email attribute mail attribute mail the attribute holding the display name of the user this will be used to greet an authenticated user display name attribute displayname the username and password of the admin user admin should be the admin username you set in the lldap configuration user uid admin ou people dc example dc com set via env var authelia authentication backend ldap user password replace me set via env var authelia authentication backend ldap password access control default policy deny rules rules applied to everyone domain public test example com policy bypass domain test example com policy one factor domain test example com policy two factor session name authelia session secret unsecure session secret set via env var authelia session secret expiration hour inactivity minutes domain example com should match whatever your root protected domain is regulation max retries find time ban time storage local path config db notifier filesystem filename config notification txt documentation here s the caddyfile caddyfile debug log format console trusted proxy list uncomment adjust the following line to configure specific ranges which should be considered as trustworthy trusted proxies authelia forward auth authelia uri api verify rd copy headers remote user remote groups remote name remote email this import needs to be included if you re relying on a trusted proxies configuration import trusted proxy list auth test example com reverse proxy authelia this import needs to be included if you re relying on a trusted proxies configuration import trusted proxy list lldap test example com reverse proxy lldap public test example com import authelia respond public test example com import authelia respond test example com import authelia respond | 1 |
228,260 | 7,548,539,559 | IssuesEvent | 2018-04-18 11:35:05 | phazonoverload/theyworkforgoldsu | https://api.github.com/repos/phazonoverload/theyworkforgoldsu | opened | Embeddable widgets | low priority | These could be users, or promises (with their respective updates). This could be used inside of existing SU sites. | 1.0 | Embeddable widgets - These could be users, or promises (with their respective updates). This could be used inside of existing SU sites. | priority | embeddable widgets these could be users or promises with their respective updates this could be used inside of existing su sites | 1 |
638,034 | 20,710,662,070 | IssuesEvent | 2022-03-12 00:17:10 | naoei/VADB-Issue-Tracker | https://api.github.com/repos/naoei/VADB-Issue-Tracker | closed | Allow admins to upload images | Priority:0 | Currently, URLs are being used for both the profile picture and the banner, the issue with this is that at times the artist might want to use another profile picture instead and as such the URL would become invalid. Letting the admins upload images themselves would circumvent this problem, however, I'm not sure if I should keep the URL text input still available and have it fetch and download the image at that time.
#8 Would also benefit from this. | 1.0 | Allow admins to upload images - Currently, URLs are being used for both the profile picture and the banner, the issue with this is that at times the artist might want to use another profile picture instead and as such the URL would become invalid. Letting the admins upload images themselves would circumvent this problem, however, I'm not sure if I should keep the URL text input still available and have it fetch and download the image at that time.
#8 Would also benefit from this. | priority | allow admins to upload images currently urls are being used for both the profile picture and the banner the issue with this is that at times the artist might want to use another profile picture instead and as such the url would become invalid letting the admins upload images themselves would circumvent this problem however i m not sure if i should keep the url text input still available and have it fetch and download the image at that time would also benefit from this | 1 |
36,530 | 9,819,127,795 | IssuesEvent | 2019-06-13 21:05:05 | cortex-lab/phy | https://api.github.com/repos/cortex-lab/phy | closed | Check for correct minimum versions of all packages on import | build/tests | Or at least main functions.
I don't know what the best way to do this would be? But we should make sure that, at least, compatible versions of most of the libraries we use are imported.
AKA it would be much more helpful to fail with:
`ERROR: numpy 1.7 is installed, minimum required version 1.8. Please upgrade`
than
`np has no attribute "full_like"` or whatever
| 1.0 | Check for correct minimum versions of all packages on import - Or at least main functions.
I don't know what the best way to do this would be? But we should make sure that, at least, compatible versions of most of the libraries we use are imported.
AKA it would be much more helpful to fail with:
`ERROR: numpy 1.7 is installed, minimum required version 1.8. Please upgrade`
than
`np has no attribute "full_like"` or whatever
| non_priority | check for correct minimum versions of all packages on import or at least main functions i don t know what the best way to do this would be but we should make sure that at least compatible versions of most of the libraries we use are imported aka it would be much more helpful to fail with error numpy is installed minimum required version please upgrade than np has no attribute full like or whatever | 0 |
318,814 | 9,702,556,446 | IssuesEvent | 2019-05-27 09:03:12 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | m.youtube.com - see bug description | browser-focus-geckoview engine-gecko priority-critical | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://m.youtube.com/watch?v=VSRnFRHWthA
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: video is green and ghost-in
**Steps to Reproduce**:
Just used Firefox focus to watch YouTube, this issue doesn't exist with the YouTube app.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | m.youtube.com - see bug description - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://m.youtube.com/watch?v=VSRnFRHWthA
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: video is green and ghost-in
**Steps to Reproduce**:
Just used Firefox focus to watch YouTube, this issue doesn't exist with the YouTube app.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | m youtube com see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description video is green and ghost in steps to reproduce just used firefox focus to watch youtube this issue doesn t exist with the youtube app browser configuration none from with ❤️ | 1 |
495,821 | 14,288,707,781 | IssuesEvent | 2020-11-23 18:06:15 | ntop/ntopng | https://api.github.com/repos/ntop/ntopng | closed | Link Redirection HS | low-priority bug | Dear Ntopng team,
please find in atachments a broken displayed link on the word heuristic ( Direct link : https://resources.sei.cmu.edu/asset_files/Presentation/2010_017_...)

Regards | 1.0 | Link Redirection HS - Dear Ntopng team,
please find in atachments a broken displayed link on the word heuristic ( Direct link : https://resources.sei.cmu.edu/asset_files/Presentation/2010_017_...)

Regards | priority | link redirection hs dear ntopng team please find in atachments a broken displayed link on the word heuristic direct link regards | 1 |
177,349 | 14,624,907,542 | IssuesEvent | 2020-12-23 07:25:19 | database-rider/database-rider | https://api.github.com/repos/database-rider/database-rider | closed | Adds Scriptable dataset docs to readme | documentation | Scriptable dataset docs are only available in [living docs](https://database-rider.github.io/database-rider/1.2.10/documentation.html#Dynamic-data-using-scritable-datasets). Also add it to readme. | 1.0 | Adds Scriptable dataset docs to readme - Scriptable dataset docs are only available in [living docs](https://database-rider.github.io/database-rider/1.2.10/documentation.html#Dynamic-data-using-scritable-datasets). Also add it to readme. | non_priority | adds scriptable dataset docs to readme scriptable dataset docs are only available in also add it to readme | 0 |
702,934 | 24,141,584,021 | IssuesEvent | 2022-09-21 15:12:32 | easystats/insight | https://api.github.com/repos/easystats/insight | closed | Symbol "Fei" misaligned with `export_table()` | bug :bug: low priority :sleeping: 3 investigators :grey_question::question: | @mattansb @bwiernik any idea?
``` r
x <- data.frame(phi_adjusted = .3, Glass_delta = .4, Fei = .3, Epsilon2 = .7, R2 = 0.4)
insight::format_table(x, use_symbols = TRUE) |>
insight::export_table()
#> ϕ (adj.) | Glass' Δ | פ | ϵ² | R²
#> ------------------------------------------
#> 0.30 | 0.40 | 0.30 | 0.70 | 0.40
x <- data.frame(phi_adjusted = .3, Glass_delta = .4, Epsilon2 = .7, R2 = 0.4)
insight::format_table(x, use_symbols = TRUE) |>
insight::export_table()
#> ϕ (adj.) | Glass' Δ | ϵ² | R²
#> ---------------------------------
#> 0.30 | 0.40 | 0.70 | 0.40
```
<sup>Created on 2022-09-21 with [reprex v2.0.2](https://reprex.tidyverse.org)</sup>
| 1.0 | Symbol "Fei" misaligned with `export_table()` - @mattansb @bwiernik any idea?
``` r
x <- data.frame(phi_adjusted = .3, Glass_delta = .4, Fei = .3, Epsilon2 = .7, R2 = 0.4)
insight::format_table(x, use_symbols = TRUE) |>
insight::export_table()
#> ϕ (adj.) | Glass' Δ | פ | ϵ² | R²
#> ------------------------------------------
#> 0.30 | 0.40 | 0.30 | 0.70 | 0.40
x <- data.frame(phi_adjusted = .3, Glass_delta = .4, Epsilon2 = .7, R2 = 0.4)
insight::format_table(x, use_symbols = TRUE) |>
insight::export_table()
#> ϕ (adj.) | Glass' Δ | ϵ² | R²
#> ---------------------------------
#> 0.30 | 0.40 | 0.70 | 0.40
```
<sup>Created on 2022-09-21 with [reprex v2.0.2](https://reprex.tidyverse.org)</sup>
| priority | symbol fei misaligned with export table mattansb bwiernik any idea r x data frame phi adjusted glass delta fei insight format table x use symbols true insight export table ϕ adj glass δ פ ϵ² r² x data frame phi adjusted glass delta insight format table x use symbols true insight export table ϕ adj glass δ ϵ² r² created on with | 1 |
244,930 | 7,880,699,943 | IssuesEvent | 2018-06-26 16:40:31 | aowen87/FOO | https://api.github.com/repos/aowen87/FOO | closed | export options are confusing for vtk | Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 2 - Minor Irritation Support Group: Any Target Version: 2.13.0 bug version: 2.10.0 | see attached picture of the export options for vtk.
two obvious issues:
* are xml and binary exclusive of each other?
* how does some one know if you check nothing you get ASCII style output?
| 1.0 | export options are confusing for vtk - see attached picture of the export options for vtk.
two obvious issues:
* are xml and binary exclusive of each other?
* how does some one know if you check nothing you get ASCII style output?
| priority | export options are confusing for vtk see attached picture of the export options for vtk two obvious issues are xml and binary exclusive of each other how does some one know if you check nothing you get ascii style output | 1 |
88,204 | 15,800,749,074 | IssuesEvent | 2021-04-03 01:06:58 | hammondjm/sql | https://api.github.com/repos/hammondjm/sql | opened | CVE-2021-20190 (High) detected in jackson-databind-2.9.7.jar | security vulnerability | ## CVE-2021-20190 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: sql/sql-jdbc/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar,canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.7.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in jackson-databind before 2.9.10.7. FasterXML mishandles the interaction between serialization gadgets and typing. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability.
<p>Publish Date: 2021-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20190>CVE-2021-20190</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2854">https://github.com/FasterXML/jackson-databind/issues/2854</a></p>
<p>Release Date: 2021-01-19</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind-2.9.10.7</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.7","packageFilePaths":["/sql-jdbc/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind-2.9.10.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-20190","vulnerabilityDetails":"A flaw was found in jackson-databind before 2.9.10.7. FasterXML mishandles the interaction between serialization gadgets and typing. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20190","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-20190 (High) detected in jackson-databind-2.9.7.jar - ## CVE-2021-20190 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: sql/sql-jdbc/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar,canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.7.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in jackson-databind before 2.9.10.7. FasterXML mishandles the interaction between serialization gadgets and typing. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability.
<p>Publish Date: 2021-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20190>CVE-2021-20190</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2854">https://github.com/FasterXML/jackson-databind/issues/2854</a></p>
<p>Release Date: 2021-01-19</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind-2.9.10.7</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.7","packageFilePaths":["/sql-jdbc/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind-2.9.10.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-20190","vulnerabilityDetails":"A flaw was found in jackson-databind before 2.9.10.7. FasterXML mishandles the interaction between serialization gadgets and typing. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20190","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file sql sql jdbc build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar canner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in base branch master vulnerability details a flaw was found in jackson databind before fasterxml mishandles the interaction between serialization gadgets and typing the highest threat from this vulnerability is to data confidentiality and integrity as well as system availability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was found in jackson databind before fasterxml mishandles the interaction between serialization gadgets and typing the highest threat from this vulnerability is to data confidentiality and integrity as well as system availability vulnerabilityurl | 0 |
11,710 | 3,016,316,903 | IssuesEvent | 2015-07-30 01:21:20 | docker/swarm | https://api.github.com/repos/docker/swarm | closed | Fix race conditions by improving the refresh management | area/API kind/bug status/1-design-review | Right now we have a bunch of race conditions in container and image state.
For instance:
```
docker start foo
docker ps
```
- `start` will proxy the start request to the node then return to the client.
- The engine will emit a `start` event, which will trigger a refresh by the Engine
- In between steps 1 and 2, if a client emits a `ps` at the right time, it will see `foo` as stopped even though it's running
To mitigate that, we could add a `Container.Refresh()` function (which in turn would call `c.Engine.refreshContainer(c.id)`. That function would get called by every proxy that makes write calls to the engine (start, stop, ...) but not read calls (inspect). | 1.0 | Fix race conditions by improving the refresh management - Right now we have a bunch of race conditions in container and image state.
For instance:
```
docker start foo
docker ps
```
- `start` will proxy the start request to the node then return to the client.
- The engine will emit a `start` event, which will trigger a refresh by the Engine
- In between steps 1 and 2, if a client emits a `ps` at the right time, it will see `foo` as stopped even though it's running
To mitigate that, we could add a `Container.Refresh()` function (which in turn would call `c.Engine.refreshContainer(c.id)`. That function would get called by every proxy that makes write calls to the engine (start, stop, ...) but not read calls (inspect). | non_priority | fix race conditions by improving the refresh management right now we have a bunch of race conditions in container and image state for instance docker start foo docker ps start will proxy the start request to the node then return to the client the engine will emit a start event which will trigger a refresh by the engine in between steps and if a client emits a ps at the right time it will see foo as stopped even though it s running to mitigate that we could add a container refresh function which in turn would call c engine refreshcontainer c id that function would get called by every proxy that makes write calls to the engine start stop but not read calls inspect | 0 |
40,274 | 2,868,315,567 | IssuesEvent | 2015-06-05 18:06:50 | GoogleCloudPlatform/kubernetes | https://api.github.com/repos/GoogleCloudPlatform/kubernetes | closed | Demo service account to replace kubernetes-ro use cases in demos | priority/P1 team/cluster team/cluster-lifecycle | # Problem
- Several of our examples use kubernetes-ro.
- cassandra, hazelcast, and rethinkdb examples use it for getting endpoints
- phabricator examples uses it for getting pods
- other examples may use it, pending further testing
- probably other config not in our repo probably uses it.
- We have use cases where nodes/pods need to reach the master over the internet, rather than over a private network. In those cases, we want to disable the kubernetes-ro service. Otherwise, it leaves cluster information exposed to the world.
# Options
1. disable kubernetes-ro and let examples be broken on those types of clusters described above.
- problem: breaks examples, and provides inconsistent cluster semantics for config authors to deal with.
1. something on the node proxies the kubernetes-ro service from the pod over ssl-with-auth to the master.
- problem: introduces a mechanism with apparently limited general utility
1. change all the examples on the site to use the http://kubernetes service, using an credential which allows readonly access.
- problem: We could change the ones on this site, but there may be others
- problem: It is a non-trivial amount of work to get all the examples working, as they use a variety of languages and libraries, which may not support https with token-based auth as well as our pkg/client.
Recommendation: Last option. It is a uses an existing, general mechanism (secrets) and keeps all clusters working the same.
# Proposal
- Add a "readonly-user"
- adds a service account token/name at cluster startup, in the same way other service accounts have been added.
- adds a secret containing this token.
- Set policy for this service account to allow reading (events, endpoints, pods, services, rcs)
- Merge https://github.com/GoogleCloudPlatform/kubernetes/pull/7101 by @liggitt
- For each example currently using kubernetes-ro:
1. update each pod to set "pod.spec.serviceAccount" to "readonly-user".
- with #7101, causes pod to automatically be filled in with a secret volume containing a token for the readonly-user account.
2. update places that use http and kubernetes-ro service to use https and kubernetes service.
3. in those same places, add a token to header of the request.
4. in those same places, set https options to use insecure connection. (Or find a way to inject the master certs into the pod)
There are a number of details to sort out once there is general agreement on the general approach. | 1.0 | Demo service account to replace kubernetes-ro use cases in demos - # Problem
- Several of our examples use kubernetes-ro.
- cassandra, hazelcast, and rethinkdb examples use it for getting endpoints
- phabricator examples uses it for getting pods
- other examples may use it, pending further testing
- probably other config not in our repo probably uses it.
- We have use cases where nodes/pods need to reach the master over the internet, rather than over a private network. In those cases, we want to disable the kubernetes-ro service. Otherwise, it leaves cluster information exposed to the world.
# Options
1. disable kubernetes-ro and let examples be broken on those types of clusters described above.
- problem: breaks examples, and provides inconsistent cluster semantics for config authors to deal with.
1. something on the node proxies the kubernetes-ro service from the pod over ssl-with-auth to the master.
- problem: introduces a mechanism with apparently limited general utility
1. change all the examples on the site to use the http://kubernetes service, using an credential which allows readonly access.
- problem: We could change the ones on this site, but there may be others
- problem: It is a non-trivial amount of work to get all the examples working, as they use a variety of languages and libraries, which may not support https with token-based auth as well as our pkg/client.
Recommendation: Last option. It is a uses an existing, general mechanism (secrets) and keeps all clusters working the same.
# Proposal
- Add a "readonly-user"
- adds a service account token/name at cluster startup, in the same way other service accounts have been added.
- adds a secret containing this token.
- Set policy for this service account to allow reading (events, endpoints, pods, services, rcs)
- Merge https://github.com/GoogleCloudPlatform/kubernetes/pull/7101 by @liggitt
- For each example currently using kubernetes-ro:
1. update each pod to set "pod.spec.serviceAccount" to "readonly-user".
- with #7101, causes pod to automatically be filled in with a secret volume containing a token for the readonly-user account.
2. update places that use http and kubernetes-ro service to use https and kubernetes service.
3. in those same places, add a token to header of the request.
4. in those same places, set https options to use insecure connection. (Or find a way to inject the master certs into the pod)
There are a number of details to sort out once there is general agreement on the general approach. | priority | demo service account to replace kubernetes ro use cases in demos problem several of our examples use kubernetes ro cassandra hazelcast and rethinkdb examples use it for getting endpoints phabricator examples uses it for getting pods other examples may use it pending further testing probably other config not in our repo probably uses it we have use cases where nodes pods need to reach the master over the internet rather than over a private network in those cases we want to disable the kubernetes ro service otherwise it leaves cluster information exposed to the world options disable kubernetes ro and let examples be broken on those types of clusters described above problem breaks examples and provides inconsistent cluster semantics for config authors to deal with something on the node proxies the kubernetes ro service from the pod over ssl with auth to the master problem introduces a mechanism with apparently limited general utility change all the examples on the site to use the service using an credential which allows readonly access problem we could change the ones on this site but there may be others problem it is a non trivial amount of work to get all the examples working as they use a variety of languages and libraries which may not support https with token based auth as well as our pkg client recommendation last option it is a uses an existing general mechanism secrets and keeps all clusters working the same proposal add a readonly user adds a service account token name at cluster startup in the same way other service accounts have been added adds a secret containing this token set policy for this service account to allow reading events endpoints pods services rcs merge by liggitt for each example currently using kubernetes ro update each pod to set pod spec serviceaccount to readonly user with causes pod to automatically be filled in with a secret volume containing a token for the readonly user account update places that use http and kubernetes ro service to use https and kubernetes service in those same places add a token to header of the request in those same places set https options to use insecure connection or find a way to inject the master certs into the pod there are a number of details to sort out once there is general agreement on the general approach | 1 |
654,627 | 21,658,278,872 | IssuesEvent | 2022-05-06 16:13:30 | onicagroup/runway | https://api.github.com/repos/onicagroup/runway | opened | [BUG] Getting error message runway fresh install TypeError: '<' not supported between instances of 'str' and 'int' | bug priority:low status:review_needed | ### Bug Description
Running runway for the first time on a fresh Amazon Linux2 machine, or on MacOS gives the following error:
runway] using backend values from runway.yml
Traceback (most recent call last):
File "runway-script.py", line 2, in <module>
File "click/core.py", line 1128, in __call__
File "click/core.py", line 1053, in main
File "runway/_cli/main.py", line 33, in invoke
File "click/core.py", line 1659, in invoke
File "click/core.py", line 1395, in invoke
File "click/core.py", line 754, in invoke
File "click/decorators.py", line 26, in new_func
File "runway/_cli/commands/_plan.py", line 47, in plan
File "runway/core/__init__.py", line 130, in plan
File "runway/core/__init__.py", line 260, in __run_action
File "runway/core/components/_deployment.py", line 313, in run_list
File "runway/core/components/_deployment.py", line 163, in plan
File "runway/core/components/_deployment.py", line 273, in __sync
File "runway/core/components/_deployment.py", line 188, in run
File "runway/core/components/_module.py", line 325, in run_list
File "runway/core/components/_module.py", line 212, in plan
File "runway/core/components/_module.py", line 242, in run
File "runway/module/terraform.py", line 354, in plan
File "runway/module/terraform.py", line 552, in run
File "runway/module/terraform.py", line 444, in terraform_init
File "runway/module/terraform.py", line 259, in gen_command
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/module/terraform.py", line 186, in tf_bin
File "runway/env_mgr/tfenv.py", line 377, in install
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/env_mgr/tfenv.py", line 307, in version
File "runway/env_mgr/tfenv.py", line 119, in get_available_tf_versions
File "distutils/version.py", line 52, in __lt__
File "distutils/version.py", line 337, in _cmp
TypeError: '<' not supported between instances of 'str' and 'int'
[2710] Failed to execute script 'runway-script' due to unhandled exception!
Runway version 2.6.1 (also tried with 2.4.4)
Same repo with runway configuration and same runway version run successfully on other machines where runway was running in the past
### Expected Behavior
Successful run or runway
### Steps To Reproduce
- Spin up a new Linux Amazon2 machine with admin IAM policy/role
- sudo yum update -y
- installing runway using curl, adding execute permissions and moving to /usr/bin
- install git
- clone the repo locally
- executing ENVIRONMENT_DEPLOY=dev runway plan.
- ...
- Getting an error:
[runway]
[runway]
[runway] XXXXXXX:processing deployment (in progress)
[runway] XXXXXXX:processing regions sequentially...
[runway] XXXXXXX:verified current AWS account matches required account id "XXXXXXXX"
[runway]
[runway] XXXXXX.XXXXXXX:processing module in us-east-1 (in progress)
[runway] XXXX-group-2-internal-lb:init (in progress)
[runway] using backend values from runway.yml
Traceback (most recent call last):
File "runway-script.py", line 2, in <module>
File "click/core.py", line 1128, in __call__
File "click/core.py", line 1053, in main
File "runway/_cli/main.py", line 33, in invoke
File "click/core.py", line 1659, in invoke
File "click/core.py", line 1395, in invoke
File "click/core.py", line 754, in invoke
File "click/decorators.py", line 26, in new_func
File "runway/_cli/commands/_plan.py", line 47, in plan
File "runway/core/__init__.py", line 130, in plan
File "runway/core/__init__.py", line 260, in __run_action
File "runway/core/components/_deployment.py", line 313, in run_list
File "runway/core/components/_deployment.py", line 163, in plan
File "runway/core/components/_deployment.py", line 273, in __sync
File "runway/core/components/_deployment.py", line 188, in run
File "runway/core/components/_module.py", line 325, in run_list
File "runway/core/components/_module.py", line 212, in plan
File "runway/core/components/_module.py", line 242, in run
File "runway/module/terraform.py", line 354, in plan
File "runway/module/terraform.py", line 552, in run
File "runway/module/terraform.py", line 444, in terraform_init
File "runway/module/terraform.py", line 259, in gen_command
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/module/terraform.py", line 186, in tf_bin
File "runway/env_mgr/tfenv.py", line 377, in install
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/env_mgr/tfenv.py", line 307, in version
File "runway/env_mgr/tfenv.py", line 119, in get_available_tf_versions
File "distutils/version.py", line 52, in __lt__
File "distutils/version.py", line 337, in _cmp
TypeError: '<' not supported between instances of 'str' and 'int'
[2710] Failed to execute script 'runway-script' due to unhandled exception!
### Runway version
2.6.1 and 2.4.4
### Installation Type
direct download (curl, wget, etc)
### OS / Environment
On not working machines
- OS: Amazon Linux 2
- python2 version: 2.7.18
- python3 version: 3.7.10
- terraform version: 1.1.0
On working machine
- python2 version: 2.7.17
- python3 version: 3.8.2
- terraform version: 1.1.0
### Anything else?
_No response_ | 1.0 | [BUG] Getting error message runway fresh install TypeError: '<' not supported between instances of 'str' and 'int' - ### Bug Description
Running runway for the first time on a fresh Amazon Linux2 machine, or on MacOS gives the following error:
runway] using backend values from runway.yml
Traceback (most recent call last):
File "runway-script.py", line 2, in <module>
File "click/core.py", line 1128, in __call__
File "click/core.py", line 1053, in main
File "runway/_cli/main.py", line 33, in invoke
File "click/core.py", line 1659, in invoke
File "click/core.py", line 1395, in invoke
File "click/core.py", line 754, in invoke
File "click/decorators.py", line 26, in new_func
File "runway/_cli/commands/_plan.py", line 47, in plan
File "runway/core/__init__.py", line 130, in plan
File "runway/core/__init__.py", line 260, in __run_action
File "runway/core/components/_deployment.py", line 313, in run_list
File "runway/core/components/_deployment.py", line 163, in plan
File "runway/core/components/_deployment.py", line 273, in __sync
File "runway/core/components/_deployment.py", line 188, in run
File "runway/core/components/_module.py", line 325, in run_list
File "runway/core/components/_module.py", line 212, in plan
File "runway/core/components/_module.py", line 242, in run
File "runway/module/terraform.py", line 354, in plan
File "runway/module/terraform.py", line 552, in run
File "runway/module/terraform.py", line 444, in terraform_init
File "runway/module/terraform.py", line 259, in gen_command
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/module/terraform.py", line 186, in tf_bin
File "runway/env_mgr/tfenv.py", line 377, in install
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/env_mgr/tfenv.py", line 307, in version
File "runway/env_mgr/tfenv.py", line 119, in get_available_tf_versions
File "distutils/version.py", line 52, in __lt__
File "distutils/version.py", line 337, in _cmp
TypeError: '<' not supported between instances of 'str' and 'int'
[2710] Failed to execute script 'runway-script' due to unhandled exception!
Runway version 2.6.1 (also tried with 2.4.4)
Same repo with runway configuration and same runway version run successfully on other machines where runway was running in the past
### Expected Behavior
Successful run or runway
### Steps To Reproduce
- Spin up a new Linux Amazon2 machine with admin IAM policy/role
- sudo yum update -y
- installing runway using curl, adding execute permissions and moving to /usr/bin
- install git
- clone the repo locally
- executing ENVIRONMENT_DEPLOY=dev runway plan.
- ...
- Getting an error:
[runway]
[runway]
[runway] XXXXXXX:processing deployment (in progress)
[runway] XXXXXXX:processing regions sequentially...
[runway] XXXXXXX:verified current AWS account matches required account id "XXXXXXXX"
[runway]
[runway] XXXXXX.XXXXXXX:processing module in us-east-1 (in progress)
[runway] XXXX-group-2-internal-lb:init (in progress)
[runway] using backend values from runway.yml
Traceback (most recent call last):
File "runway-script.py", line 2, in <module>
File "click/core.py", line 1128, in __call__
File "click/core.py", line 1053, in main
File "runway/_cli/main.py", line 33, in invoke
File "click/core.py", line 1659, in invoke
File "click/core.py", line 1395, in invoke
File "click/core.py", line 754, in invoke
File "click/decorators.py", line 26, in new_func
File "runway/_cli/commands/_plan.py", line 47, in plan
File "runway/core/__init__.py", line 130, in plan
File "runway/core/__init__.py", line 260, in __run_action
File "runway/core/components/_deployment.py", line 313, in run_list
File "runway/core/components/_deployment.py", line 163, in plan
File "runway/core/components/_deployment.py", line 273, in __sync
File "runway/core/components/_deployment.py", line 188, in run
File "runway/core/components/_module.py", line 325, in run_list
File "runway/core/components/_module.py", line 212, in plan
File "runway/core/components/_module.py", line 242, in run
File "runway/module/terraform.py", line 354, in plan
File "runway/module/terraform.py", line 552, in run
File "runway/module/terraform.py", line 444, in terraform_init
File "runway/module/terraform.py", line 259, in gen_command
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/module/terraform.py", line 186, in tf_bin
File "runway/env_mgr/tfenv.py", line 377, in install
File "backports/cached_property/__init__.py", line 80, in __get__
File "runway/env_mgr/tfenv.py", line 307, in version
File "runway/env_mgr/tfenv.py", line 119, in get_available_tf_versions
File "distutils/version.py", line 52, in __lt__
File "distutils/version.py", line 337, in _cmp
TypeError: '<' not supported between instances of 'str' and 'int'
[2710] Failed to execute script 'runway-script' due to unhandled exception!
### Runway version
2.6.1 and 2.4.4
### Installation Type
direct download (curl, wget, etc)
### OS / Environment
On not working machines
- OS: Amazon Linux 2
- python2 version: 2.7.18
- python3 version: 3.7.10
- terraform version: 1.1.0
On working machine
- python2 version: 2.7.17
- python3 version: 3.8.2
- terraform version: 1.1.0
### Anything else?
_No response_ | priority | getting error message runway fresh install typeerror not supported between instances of str and int bug description running runway for the first time on a fresh amazon machine or on macos gives the following error runway using backend values from runway yml traceback most recent call last file runway script py line in file click core py line in call file click core py line in main file runway cli main py line in invoke file click core py line in invoke file click core py line in invoke file click core py line in invoke file click decorators py line in new func file runway cli commands plan py line in plan file runway core init py line in plan file runway core init py line in run action file runway core components deployment py line in run list file runway core components deployment py line in plan file runway core components deployment py line in sync file runway core components deployment py line in run file runway core components module py line in run list file runway core components module py line in plan file runway core components module py line in run file runway module terraform py line in plan file runway module terraform py line in run file runway module terraform py line in terraform init file runway module terraform py line in gen command file backports cached property init py line in get file runway module terraform py line in tf bin file runway env mgr tfenv py line in install file backports cached property init py line in get file runway env mgr tfenv py line in version file runway env mgr tfenv py line in get available tf versions file distutils version py line in lt file distutils version py line in cmp typeerror not supported between instances of str and int failed to execute script runway script due to unhandled exception runway version also tried with same repo with runway configuration and same runway version run successfully on other machines where runway was running in the past expected behavior successful run or runway steps to reproduce spin up a new linux machine with admin iam policy role sudo yum update y installing runway using curl adding execute permissions and moving to usr bin install git clone the repo locally executing environment deploy dev runway plan getting an error xxxxxxx processing deployment in progress xxxxxxx processing regions sequentially xxxxxxx verified current aws account matches required account id xxxxxxxx xxxxxx xxxxxxx processing module in us east in progress xxxx group internal lb init in progress using backend values from runway yml traceback most recent call last file runway script py line in file click core py line in call file click core py line in main file runway cli main py line in invoke file click core py line in invoke file click core py line in invoke file click core py line in invoke file click decorators py line in new func file runway cli commands plan py line in plan file runway core init py line in plan file runway core init py line in run action file runway core components deployment py line in run list file runway core components deployment py line in plan file runway core components deployment py line in sync file runway core components deployment py line in run file runway core components module py line in run list file runway core components module py line in plan file runway core components module py line in run file runway module terraform py line in plan file runway module terraform py line in run file runway module terraform py line in terraform init file runway module terraform py line in gen command file backports cached property init py line in get file runway module terraform py line in tf bin file runway env mgr tfenv py line in install file backports cached property init py line in get file runway env mgr tfenv py line in version file runway env mgr tfenv py line in get available tf versions file distutils version py line in lt file distutils version py line in cmp typeerror not supported between instances of str and int failed to execute script runway script due to unhandled exception runway version and installation type direct download curl wget etc os environment on not working machines os amazon linux version version terraform version on working machine version version terraform version anything else no response | 1 |
81,414 | 15,726,503,788 | IssuesEvent | 2021-03-29 11:24:31 | parallaxsecond/parsec | https://api.github.com/repos/parallaxsecond/parsec | opened | Log when skipping test | code health testing | We have a lot of end-to-end tests that depend on specific operations being supported by the providers. Many of the tests check if the operation is present for the current provider under test, and if it isn't, the test returns and ends successfully. I think we should add a log message before returning, something like `Operation PsaSignHash is not supported by provider TpmProvider, skipping test.` (the test name isn't necessary since I think it is included in the log prefix). This way it's easier to see which tests were skipped if we need to investigate some logs. | 1.0 | Log when skipping test - We have a lot of end-to-end tests that depend on specific operations being supported by the providers. Many of the tests check if the operation is present for the current provider under test, and if it isn't, the test returns and ends successfully. I think we should add a log message before returning, something like `Operation PsaSignHash is not supported by provider TpmProvider, skipping test.` (the test name isn't necessary since I think it is included in the log prefix). This way it's easier to see which tests were skipped if we need to investigate some logs. | non_priority | log when skipping test we have a lot of end to end tests that depend on specific operations being supported by the providers many of the tests check if the operation is present for the current provider under test and if it isn t the test returns and ends successfully i think we should add a log message before returning something like operation psasignhash is not supported by provider tpmprovider skipping test the test name isn t necessary since i think it is included in the log prefix this way it s easier to see which tests were skipped if we need to investigate some logs | 0 |
116,418 | 14,958,976,397 | IssuesEvent | 2021-01-27 02:02:48 | dotnet/winforms | https://api.github.com/repos/dotnet/winforms | closed | Button component not available in Toolbox | area: VS designer | <!-- Read https://github.com/dotnet/winforms/blob/master/Documentation/issue-guide.md -->
* .NET Core Version:
3.1 release in VS 16.6
* Have you experienced this same bug with .NET Framework?:
No
**Problem description:**
The Button component is missing from the VS Toolbox... Label (and many others are there), but no Button.
**Expected behavior:**
The Button component would be available (visible) in the Toolbox.
**Minimal repro:**
Other than it is happening "on my machine", I'm not sure how to repro - I'm also not aware of any way to *reset* the Toolbox.
(NOTE: I am working with WinForms in Visual Basic... so not sure if this might be related.) | 1.0 | Button component not available in Toolbox - <!-- Read https://github.com/dotnet/winforms/blob/master/Documentation/issue-guide.md -->
* .NET Core Version:
3.1 release in VS 16.6
* Have you experienced this same bug with .NET Framework?:
No
**Problem description:**
The Button component is missing from the VS Toolbox... Label (and many others are there), but no Button.
**Expected behavior:**
The Button component would be available (visible) in the Toolbox.
**Minimal repro:**
Other than it is happening "on my machine", I'm not sure how to repro - I'm also not aware of any way to *reset* the Toolbox.
(NOTE: I am working with WinForms in Visual Basic... so not sure if this might be related.) | non_priority | button component not available in toolbox net core version release in vs have you experienced this same bug with net framework no problem description the button component is missing from the vs toolbox label and many others are there but no button expected behavior the button component would be available visible in the toolbox minimal repro other than it is happening on my machine i m not sure how to repro i m also not aware of any way to reset the toolbox note i am working with winforms in visual basic so not sure if this might be related | 0 |
282,326 | 24,468,723,685 | IssuesEvent | 2022-10-07 17:29:55 | IntellectualSites/PlotSquared | https://api.github.com/repos/IntellectualSites/PlotSquared | opened | Nether-Portal BUG | Requires Testing | ### Server Implementation
Paper
### Server Version
1.19.2
### Describe the bug
My friend used the nether portal on his plot. Then I noticed the error (see attachment) in my console. I can't say for sure if it was when he enters the nether portal on his plot OR if the error came after he used the ALREADY EXISTING portal inside the nether to get back to the normal world again (when he used his portal on his plot he did spawn inside the already exisiting portal inside the nether). But I can say for sure that the error must have a connection to the nether portal on his plot.
### To Reproduce
1. MAKE SURE you have already a manual / regular created portal from the normal world to the nether
2. Now create a nether portal on your plot & go trough it
3. use (an already existing!) portal inside the nether to get back to the normal world
4. The error should appear in the console
### Expected behaviour
No error should appear.
### Screenshots / Videos


### Error log (if applicable)
https://paste.gg/p/anonymous/71dca8345f304669b04c63249bd3e6b9
### Plot Debugpaste
It said "Failed to create the debugpaste: request must contain a file list" so "Pierre" from Discord told me to paste the debug this time
### PlotSquared Version
6.10.0-Premium
### Checklist
- [X] I have included a Plot debugpaste.
- [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists.
### Anything else?
_No response_ | 1.0 | Nether-Portal BUG - ### Server Implementation
Paper
### Server Version
1.19.2
### Describe the bug
My friend used the nether portal on his plot. Then I noticed the error (see attachment) in my console. I can't say for sure if it was when he enters the nether portal on his plot OR if the error came after he used the ALREADY EXISTING portal inside the nether to get back to the normal world again (when he used his portal on his plot he did spawn inside the already exisiting portal inside the nether). But I can say for sure that the error must have a connection to the nether portal on his plot.
### To Reproduce
1. MAKE SURE you have already a manual / regular created portal from the normal world to the nether
2. Now create a nether portal on your plot & go trough it
3. use (an already existing!) portal inside the nether to get back to the normal world
4. The error should appear in the console
### Expected behaviour
No error should appear.
### Screenshots / Videos


### Error log (if applicable)
https://paste.gg/p/anonymous/71dca8345f304669b04c63249bd3e6b9
### Plot Debugpaste
It said "Failed to create the debugpaste: request must contain a file list" so "Pierre" from Discord told me to paste the debug this time
### PlotSquared Version
6.10.0-Premium
### Checklist
- [X] I have included a Plot debugpaste.
- [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists.
### Anything else?
_No response_ | non_priority | nether portal bug server implementation paper server version describe the bug my friend used the nether portal on his plot then i noticed the error see attachment in my console i can t say for sure if it was when he enters the nether portal on his plot or if the error came after he used the already existing portal inside the nether to get back to the normal world again when he used his portal on his plot he did spawn inside the already exisiting portal inside the nether but i can say for sure that the error must have a connection to the nether portal on his plot to reproduce make sure you have already a manual regular created portal from the normal world to the nether now create a nether portal on your plot go trough it use an already existing portal inside the nether to get back to the normal world the error should appear in the console expected behaviour no error should appear screenshots videos error log if applicable plot debugpaste it said failed to create the debugpaste request must contain a file list so pierre from discord told me to paste the debug this time plotsquared version premium checklist i have included a plot debugpaste i am using the newest build from and the issue still persists anything else no response | 0 |
776,178 | 27,250,013,739 | IssuesEvent | 2023-02-22 07:11:53 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.reuters.com - site is not usable | priority-important browser-focus-geckoview engine-gecko bugbug-reopened | <!-- @browser: Firefox Mobile 110.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:109.0) Gecko/110.0 Firefox/110.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/118531 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.reuters.com/world/europe/biden-discuss-more-troops-upcoming-visit-polish-pm-says-2023-02-19/
**Browser / Version**: Firefox Mobile 110.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Open link, Article not loaded / shown
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230213213738</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/2/206a5da3-08ff-4ec9-9f7a-2250d194e62f)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.reuters.com - site is not usable - <!-- @browser: Firefox Mobile 110.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:109.0) Gecko/110.0 Firefox/110.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/118531 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.reuters.com/world/europe/biden-discuss-more-troops-upcoming-visit-polish-pm-says-2023-02-19/
**Browser / Version**: Firefox Mobile 110.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Open link, Article not loaded / shown
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230213213738</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/2/206a5da3-08ff-4ec9-9f7a-2250d194e62f)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce open link article not loaded shown browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
801,953 | 28,508,165,664 | IssuesEvent | 2023-04-19 00:14:37 | bcgov/platform-services-registry | https://api.github.com/repos/bcgov/platform-services-registry | closed | Gather API requirements for the public cloud team | Sprint Goal priority/ high *team/ app development* | **Describe the issue**
Public cloud will integrate with our platform and will use our API. We may need to build out new API endpoints or show them our existing endpoints.
We need to kick this off with a Technical meeting with the public cloud.
These people should be in attendance:
- Sal
- Artem
- Zhanna
- Oamar
- Warren
- Carles
- Prabhu
Optional attendees:
- Billy (as he has been working with public cloud API)
- Nick (Platform security)
- Ian
- Shelly
In preparation for the meeting, Oamar will:
- Meet with Billy to get updated on public cloud
- Update our API documentation and create examples for fetching and creating data on each endpoint.
- Create an example of creating an Auth token for API endpoints
**Definition of done**
- [ ] Meet with Public Cloud
- [ ] We understand the public cloud requirements for our API
| 1.0 | Gather API requirements for the public cloud team - **Describe the issue**
Public cloud will integrate with our platform and will use our API. We may need to build out new API endpoints or show them our existing endpoints.
We need to kick this off with a Technical meeting with the public cloud.
These people should be in attendance:
- Sal
- Artem
- Zhanna
- Oamar
- Warren
- Carles
- Prabhu
Optional attendees:
- Billy (as he has been working with public cloud API)
- Nick (Platform security)
- Ian
- Shelly
In preparation for the meeting, Oamar will:
- Meet with Billy to get updated on public cloud
- Update our API documentation and create examples for fetching and creating data on each endpoint.
- Create an example of creating an Auth token for API endpoints
**Definition of done**
- [ ] Meet with Public Cloud
- [ ] We understand the public cloud requirements for our API
| priority | gather api requirements for the public cloud team describe the issue public cloud will integrate with our platform and will use our api we may need to build out new api endpoints or show them our existing endpoints we need to kick this off with a technical meeting with the public cloud these people should be in attendance sal artem zhanna oamar warren carles prabhu optional attendees billy as he has been working with public cloud api nick platform security ian shelly in preparation for the meeting oamar will meet with billy to get updated on public cloud update our api documentation and create examples for fetching and creating data on each endpoint create an example of creating an auth token for api endpoints definition of done meet with public cloud we understand the public cloud requirements for our api | 1 |
201,534 | 7,033,093,223 | IssuesEvent | 2017-12-27 08:50:24 | Eustacio/seed-starter-manager | https://api.github.com/repos/Eustacio/seed-starter-manager | closed | Replace usages of the error-label.directive by the input-with-validation.directive | Front-end Priority: medium Type: improvement | The input-with-validation.directive was created to replace the error-label.directive, therefore we need replace all usages and then delete it.
The error-label.directive is currently used in the following places:
- [ ] seed-variety.component.html
- [ ] feature.component.html
- [ ] material-type.component.html | 1.0 | Replace usages of the error-label.directive by the input-with-validation.directive - The input-with-validation.directive was created to replace the error-label.directive, therefore we need replace all usages and then delete it.
The error-label.directive is currently used in the following places:
- [ ] seed-variety.component.html
- [ ] feature.component.html
- [ ] material-type.component.html | priority | replace usages of the error label directive by the input with validation directive the input with validation directive was created to replace the error label directive therefore we need replace all usages and then delete it the error label directive is currently used in the following places seed variety component html feature component html material type component html | 1 |
554,744 | 16,437,303,304 | IssuesEvent | 2021-05-20 10:40:57 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID :219588] Out-of-bounds access in tests/bluetooth/tester/src/gatt.c | Coverity area: Bluetooth bug priority: low |
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/bluetooth/tester/src/gatt.c
Category: Memory - corruptions
Function: `alloc_value`
Component: Tests
CID: [219588](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219588)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/bluetooth/tester/src/gatt.c
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| 1.0 | [Coverity CID :219588] Out-of-bounds access in tests/bluetooth/tester/src/gatt.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/bluetooth/tester/src/gatt.c
Category: Memory - corruptions
Function: `alloc_value`
Component: Tests
CID: [219588](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219588)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/bluetooth/tester/src/gatt.c
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| priority | out of bounds access in tests bluetooth tester src gatt c static code scan issues found in file category memory corruptions function alloc value component tests cid details please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file | 1 |
130,550 | 27,720,301,363 | IssuesEvent | 2023-03-14 20:03:21 | microsoft/vscode-cpptools | https://api.github.com/repos/microsoft/vscode-cpptools | closed | Clang Tidy Code analysis not working with a Workspace file. | Language Service investigate Feature: Configuration Feature: Code Analysis |
Type: <b>Bug</b>
I created a simple project with the following layout:
```text
MyProject
.vs/settings.json
MyProject
Sample
Sources
main.cpp
CMakeLists.txt
CMakeLists.txt
.clang-tidy
MyProject.code-workspace
```
I tried moving .clang-tidy around. I tried setting:
```json
"C_Cpp.codeAnalysis.clangTidy.enabled": true,
```
in `.vscode/settings.json` and also in `MyProject.code-workspace` but I couldn't get the clang-tidy code analisys to work.
In a most basic project without a workspace file it works.
Also is this presets setting supported for Clang-Tidy:
```json
"vendor": {
"microsoft.com/VisualStudioSettings/CMake/1.0": {
"enableMicrosoftCodeAnalysis": true,
}
}
```
Extension version: 1.13.2
VS Code version: Code 1.72.0 (64bbfbf67ada9953918d72e1df2f4d8e537d340e, 2022-10-04T23:20:47.539Z)
OS version: Linux x64 5.15.0-48-generic snap
Modes:
Sandboxed: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i7-7700HQ CPU @ 2.80GHz (8 x 1298)|
|GPU Status|2d_canvas: unavailable_software<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: disabled_software<br>multiple_raster_threads: enabled_on<br>opengl: disabled_off<br>rasterization: disabled_software<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: disabled_software<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: unavailable_software<br>webgl2: unavailable_software<br>webgpu: disabled_off|
|Load (avg)|2, 2, 2|
|Memory (System)|7.42GB (1.38GB free)|
|Process Argv|--no-sandbox --force-user-env --unity-launch --crash-reporter-id 20280fc3-3c6d-43d3-9eb5-55c536b003d8|
|Screen Reader|no|
|VM|0%|
|DESKTOP_SESSION|ubuntu-wayland|
|XDG_CURRENT_DESKTOP|Unity|
|XDG_SESSION_DESKTOP|ubuntu-wayland|
|XDG_SESSION_TYPE|wayland|
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vsreu685:30147344
python383:30185418
vspor879:30202332
vspor708:30202333
vspor363:30204092
vslsvsres303:30308271
pythonvspyl392:30443607
vserr242:30382549
pythontb:30283811
vsjup518:30340749
pythonptprofiler:30281270
vsdfh931cf:30280410
vshan820:30294714
vstes263:30335439
vscorecescf:30445987
pythondataviewer:30285071
vscod805cf:30301675
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
cmake_vspar411:30581797
vsaa593:30376534
pythonvs932:30410667
cppdebug:30492333
vscaat:30438848
vsclangdf:30486550
c4g48928:30535728
dsvsc012cf:30540253
azure-dev_surveyone:30548225
2144e591:30553903
vsccc:30566497
pyindex848cf:30577861
nodejswelcome1:30583747
40g7c324:30573242
```
</details>
<!-- generated by issue reporter --> | 1.0 | Clang Tidy Code analysis not working with a Workspace file. -
Type: <b>Bug</b>
I created a simple project with the following layout:
```text
MyProject
.vs/settings.json
MyProject
Sample
Sources
main.cpp
CMakeLists.txt
CMakeLists.txt
.clang-tidy
MyProject.code-workspace
```
I tried moving .clang-tidy around. I tried setting:
```json
"C_Cpp.codeAnalysis.clangTidy.enabled": true,
```
in `.vscode/settings.json` and also in `MyProject.code-workspace` but I couldn't get the clang-tidy code analisys to work.
In a most basic project without a workspace file it works.
Also is this presets setting supported for Clang-Tidy:
```json
"vendor": {
"microsoft.com/VisualStudioSettings/CMake/1.0": {
"enableMicrosoftCodeAnalysis": true,
}
}
```
Extension version: 1.13.2
VS Code version: Code 1.72.0 (64bbfbf67ada9953918d72e1df2f4d8e537d340e, 2022-10-04T23:20:47.539Z)
OS version: Linux x64 5.15.0-48-generic snap
Modes:
Sandboxed: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i7-7700HQ CPU @ 2.80GHz (8 x 1298)|
|GPU Status|2d_canvas: unavailable_software<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: disabled_software<br>multiple_raster_threads: enabled_on<br>opengl: disabled_off<br>rasterization: disabled_software<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: disabled_software<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: unavailable_software<br>webgl2: unavailable_software<br>webgpu: disabled_off|
|Load (avg)|2, 2, 2|
|Memory (System)|7.42GB (1.38GB free)|
|Process Argv|--no-sandbox --force-user-env --unity-launch --crash-reporter-id 20280fc3-3c6d-43d3-9eb5-55c536b003d8|
|Screen Reader|no|
|VM|0%|
|DESKTOP_SESSION|ubuntu-wayland|
|XDG_CURRENT_DESKTOP|Unity|
|XDG_SESSION_DESKTOP|ubuntu-wayland|
|XDG_SESSION_TYPE|wayland|
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vsreu685:30147344
python383:30185418
vspor879:30202332
vspor708:30202333
vspor363:30204092
vslsvsres303:30308271
pythonvspyl392:30443607
vserr242:30382549
pythontb:30283811
vsjup518:30340749
pythonptprofiler:30281270
vsdfh931cf:30280410
vshan820:30294714
vstes263:30335439
vscorecescf:30445987
pythondataviewer:30285071
vscod805cf:30301675
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
cmake_vspar411:30581797
vsaa593:30376534
pythonvs932:30410667
cppdebug:30492333
vscaat:30438848
vsclangdf:30486550
c4g48928:30535728
dsvsc012cf:30540253
azure-dev_surveyone:30548225
2144e591:30553903
vsccc:30566497
pyindex848cf:30577861
nodejswelcome1:30583747
40g7c324:30573242
```
</details>
<!-- generated by issue reporter --> | non_priority | clang tidy code analysis not working with a workspace file type bug i created a simple project with the following layout text myproject vs settings json myproject sample sources main cpp cmakelists txt cmakelists txt clang tidy myproject code workspace i tried moving clang tidy around i tried setting json c cpp codeanalysis clangtidy enabled true in vscode settings json and also in myproject code workspace but i couldn t get the clang tidy code analisys to work in a most basic project without a workspace file it works also is this presets setting supported for clang tidy json vendor microsoft com visualstudiosettings cmake enablemicrosoftcodeanalysis true extension version vs code version code os version linux generic snap modes sandboxed no system info item value cpus intel r core tm cpu x gpu status canvas unavailable software canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing disabled software multiple raster threads enabled on opengl disabled off rasterization disabled software raw draw disabled off ok skia renderer enabled on video decode disabled software video encode disabled software vulkan disabled off webgl unavailable software unavailable software webgpu disabled off load avg memory system free process argv no sandbox force user env unity launch crash reporter id screen reader no vm desktop session ubuntu wayland xdg current desktop unity xdg session desktop ubuntu wayland xdg session type wayland a b experiments pythontb pythonptprofiler vscorecescf pythondataviewer cmake cppdebug vscaat vsclangdf azure dev surveyone vsccc | 0 |
403,801 | 11,847,092,011 | IssuesEvent | 2020-03-24 11:22:25 | dojo/widgets | https://api.github.com/repos/dojo/widgets | closed | Change TimePicker to use `Select` and convert to function based | dojo-7.0 large priority | Combobox is being deleted so timepicker should be updated to use the new approach with select and menu.
- convert to function based
- use icache
- accept initialValue
- use select
- remove combobox usage | 1.0 | Change TimePicker to use `Select` and convert to function based - Combobox is being deleted so timepicker should be updated to use the new approach with select and menu.
- convert to function based
- use icache
- accept initialValue
- use select
- remove combobox usage | priority | change timepicker to use select and convert to function based combobox is being deleted so timepicker should be updated to use the new approach with select and menu convert to function based use icache accept initialvalue use select remove combobox usage | 1 |
656,738 | 21,773,654,390 | IssuesEvent | 2022-05-13 11:40:56 | momentum-mod/game | https://api.github.com/repos/momentum-mod/game | closed | Overlapping warning boxes | Type: Bug Priority: Medium Where: Game | **Describe the bug**
The warnings for using practice mode and playing a map in a different gamemode overlap eachother
**Expected behavior**
One/both should move so that they do not overlap
**Screenshots**

**Desktop/Branch (please complete the following information):**
- Branch: 0.8.7 steam release
**Additional context**
Reported by webby in discord
| 1.0 | Overlapping warning boxes - **Describe the bug**
The warnings for using practice mode and playing a map in a different gamemode overlap eachother
**Expected behavior**
One/both should move so that they do not overlap
**Screenshots**

**Desktop/Branch (please complete the following information):**
- Branch: 0.8.7 steam release
**Additional context**
Reported by webby in discord
| priority | overlapping warning boxes describe the bug the warnings for using practice mode and playing a map in a different gamemode overlap eachother expected behavior one both should move so that they do not overlap screenshots desktop branch please complete the following information branch steam release additional context reported by webby in discord | 1 |
467,571 | 13,450,808,529 | IssuesEvent | 2020-09-08 19:09:25 | bloom-housing/bloom | https://api.github.com/repos/bloom-housing/bloom | closed | Application: Email Confirmation | High Priority in-development |
**If Lottery:**
Hello ______,
Thanks for applying. We have received your application for [Listing Name: link to listing].
Here is your confirmation number: [Confirmation Number]
What to expect next:
The lottery will be held on [Lottery Date]. Applicants will be contacted by the agent in lottery rank order until vacancies are filled.
Should your application be chosen, be prepared to fill out a more detailed application and provide required supporting documents.
If you need to update information on your application, do not apply again. Contact the agent. See below for contact information for the Agent for this listing.
[Leasing Agent Name]
[Leasing Agent Title]
[Leasing Agent Phone]
[Leasing Agent Email]
Office Hours:
[Leasing Agent Office Hours]
How are we doing? We'd like to get your [feedback: link to feedback form].
Thank you,
Alameda County - Housing and Community Development (HCD) Department
**If no Lottery and Waitlist**
Hello ______,
Thanks for applying. We have received your application for [Listing Name: link to listing].
Here is your confirmation number: [Confirmation Number]
What to expect next:
Applicants will be contacted by the agent in waitlist order until vacancies are filled.
Should your application be chosen, be prepared to fill out a more detailed application and provide required supporting documents.
If you need to update information on your application, do not apply again. Contact the agent. See below for contact information for the Agent for this listing.
[Leasing Agent Name]
[Leasing Agent Title]
[Leasing Agent Phone]
[Leasing Agent Email]
Office Hours:
[Leasing Agent Office Hours]
How are we doing? We'd like to get your [feedback: link to feedback form].
Thank you,
Alameda County - Housing and Community Development (HCD) Department
**If First Come First Serve**
Hello ______,
Thanks for applying. We have received your application for [Listing Name: link to listing].
Here is your confirmation number: [Confirmation Number]
What to expect next:
Applicants will be contacted by the property agent on a first come first serve basis until vacancies are filled.
Should your application be chosen, be prepared to fill out a more detailed application and provide required supporting documents.
If you need to update information on your application, do not apply again. Contact the agent. See below for contact information for the Agent for this listing.
[Leasing Agent Name]
[Leasing Agent Title]
[Leasing Agent Phone]
[Leasing Agent Email]
Office Hours:
[Leasing Agent Office Hours]
How are we doing? We'd like to get your [feedback: link to feedback form].
Thank you,
Alameda County - Housing and Community Development (HCD) Department | 1.0 | Application: Email Confirmation -
**If Lottery:**
Hello ______,
Thanks for applying. We have received your application for [Listing Name: link to listing].
Here is your confirmation number: [Confirmation Number]
What to expect next:
The lottery will be held on [Lottery Date]. Applicants will be contacted by the agent in lottery rank order until vacancies are filled.
Should your application be chosen, be prepared to fill out a more detailed application and provide required supporting documents.
If you need to update information on your application, do not apply again. Contact the agent. See below for contact information for the Agent for this listing.
[Leasing Agent Name]
[Leasing Agent Title]
[Leasing Agent Phone]
[Leasing Agent Email]
Office Hours:
[Leasing Agent Office Hours]
How are we doing? We'd like to get your [feedback: link to feedback form].
Thank you,
Alameda County - Housing and Community Development (HCD) Department
**If no Lottery and Waitlist**
Hello ______,
Thanks for applying. We have received your application for [Listing Name: link to listing].
Here is your confirmation number: [Confirmation Number]
What to expect next:
Applicants will be contacted by the agent in waitlist order until vacancies are filled.
Should your application be chosen, be prepared to fill out a more detailed application and provide required supporting documents.
If you need to update information on your application, do not apply again. Contact the agent. See below for contact information for the Agent for this listing.
[Leasing Agent Name]
[Leasing Agent Title]
[Leasing Agent Phone]
[Leasing Agent Email]
Office Hours:
[Leasing Agent Office Hours]
How are we doing? We'd like to get your [feedback: link to feedback form].
Thank you,
Alameda County - Housing and Community Development (HCD) Department
**If First Come First Serve**
Hello ______,
Thanks for applying. We have received your application for [Listing Name: link to listing].
Here is your confirmation number: [Confirmation Number]
What to expect next:
Applicants will be contacted by the property agent on a first come first serve basis until vacancies are filled.
Should your application be chosen, be prepared to fill out a more detailed application and provide required supporting documents.
If you need to update information on your application, do not apply again. Contact the agent. See below for contact information for the Agent for this listing.
[Leasing Agent Name]
[Leasing Agent Title]
[Leasing Agent Phone]
[Leasing Agent Email]
Office Hours:
[Leasing Agent Office Hours]
How are we doing? We'd like to get your [feedback: link to feedback form].
Thank you,
Alameda County - Housing and Community Development (HCD) Department | priority | application email confirmation if lottery hello thanks for applying we have received your application for here is your confirmation number what to expect next the lottery will be held on applicants will be contacted by the agent in lottery rank order until vacancies are filled should your application be chosen be prepared to fill out a more detailed application and provide required supporting documents if you need to update information on your application do not apply again contact the agent see below for contact information for the agent for this listing office hours how are we doing we d like to get your thank you alameda county housing and community development hcd department if no lottery and waitlist hello thanks for applying we have received your application for here is your confirmation number what to expect next applicants will be contacted by the agent in waitlist order until vacancies are filled should your application be chosen be prepared to fill out a more detailed application and provide required supporting documents if you need to update information on your application do not apply again contact the agent see below for contact information for the agent for this listing office hours how are we doing we d like to get your thank you alameda county housing and community development hcd department if first come first serve hello thanks for applying we have received your application for here is your confirmation number what to expect next applicants will be contacted by the property agent on a first come first serve basis until vacancies are filled should your application be chosen be prepared to fill out a more detailed application and provide required supporting documents if you need to update information on your application do not apply again contact the agent see below for contact information for the agent for this listing office hours how are we doing we d like to get your thank you alameda county housing and community development hcd department | 1 |
340,072 | 10,266,321,673 | IssuesEvent | 2019-08-22 21:07:05 | redhat-developer/vscode-openshift-tools | https://api.github.com/repos/redhat-developer/vscode-openshift-tools | closed | Add 'Create Service' command context menu for Projects | in progress kind/bug priority/major | Command should handle application selection the same way it is done in 'Create' command for Component:
1. `+ Create new Application` item in quick pick list along with existing applications
2. Create Application node if new Application requested | 1.0 | Add 'Create Service' command context menu for Projects - Command should handle application selection the same way it is done in 'Create' command for Component:
1. `+ Create new Application` item in quick pick list along with existing applications
2. Create Application node if new Application requested | priority | add create service command context menu for projects command should handle application selection the same way it is done in create command for component create new application item in quick pick list along with existing applications create application node if new application requested | 1 |
203,300 | 7,059,570,693 | IssuesEvent | 2018-01-05 02:34:10 | bukinoshita/taskr | https://api.github.com/repos/bukinoshita/taskr | closed | [Feature Idea] - Send task back from Today to Backlog | Priority: Medium Type: Enhancement | I believe this would be a good feature to have, in case you put as "today" a task that wasn't supposed to be for today, but you don't want to delete/recreate it.
I guess a "Backlog ↪" button or something under it whenever is on the "Today" Tab.
Will try to implement if I have the time soon, but whoever want to go ahead and do it also would be awesome. 😃
| 1.0 | [Feature Idea] - Send task back from Today to Backlog - I believe this would be a good feature to have, in case you put as "today" a task that wasn't supposed to be for today, but you don't want to delete/recreate it.
I guess a "Backlog ↪" button or something under it whenever is on the "Today" Tab.
Will try to implement if I have the time soon, but whoever want to go ahead and do it also would be awesome. 😃
| priority | send task back from today to backlog i believe this would be a good feature to have in case you put as today a task that wasn t supposed to be for today but you don t want to delete recreate it i guess a backlog ↪ button or something under it whenever is on the today tab will try to implement if i have the time soon but whoever want to go ahead and do it also would be awesome 😃 | 1 |
802,849 | 29,047,494,620 | IssuesEvent | 2023-05-13 19:03:37 | flytegg/ls-discord-bot | https://api.github.com/repos/flytegg/ls-discord-bot | opened | Reopen closing knowledgebase posts/projects | enhancement low priority | Posts seem to automatically close after X time and I think we can adjust this but not disable it within the ui. The bot should just reopen any post which closes. Closed posts are a problem because they cannot be referenced with #, so especially for suggesting knowledgebase items. | 1.0 | Reopen closing knowledgebase posts/projects - Posts seem to automatically close after X time and I think we can adjust this but not disable it within the ui. The bot should just reopen any post which closes. Closed posts are a problem because they cannot be referenced with #, so especially for suggesting knowledgebase items. | priority | reopen closing knowledgebase posts projects posts seem to automatically close after x time and i think we can adjust this but not disable it within the ui the bot should just reopen any post which closes closed posts are a problem because they cannot be referenced with so especially for suggesting knowledgebase items | 1 |
15,828 | 11,725,145,234 | IssuesEvent | 2020-03-10 12:23:34 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | Rename RuntimeOS to BuildOS | area-Infrastructure | We agreed on using the terms `BuildOS` for the build machine's OS and `TargetOS` for the target platform's OS. We should rename the existing properties to reflect that.
cc @ericstj @jkotas | 1.0 | Rename RuntimeOS to BuildOS - We agreed on using the terms `BuildOS` for the build machine's OS and `TargetOS` for the target platform's OS. We should rename the existing properties to reflect that.
cc @ericstj @jkotas | non_priority | rename runtimeos to buildos we agreed on using the terms buildos for the build machine s os and targetos for the target platform s os we should rename the existing properties to reflect that cc ericstj jkotas | 0 |
493,530 | 14,234,285,928 | IssuesEvent | 2020-11-18 13:23:40 | blockframes/blockframes | https://api.github.com/repos/blockframes/blockframes | closed | Homogenize the way users are mentioned | App - Festival 🎪 Front Medium priority Refactoring - enhancement Sanity-team | Homogenize the way users are mentioned. For now, sometimes they’re mentioned by their names (no email, no company information. In notifications for example), sometimes just their first name & company (email notifications), sometimes just by their email address (guest list).
When we mention someone, wherever this is (in a notification/an invitation/a list/ an email), it should always show first name, last name, email address & company name.
The format should be: "firstName lastName (orgName) blablabla" + adding the user email on link mailto on the firstname + lastName.
e.g. [Vincent Choukroun](Vchoukroun@cascade8.com) (Cascade8) has accepted your invitation to screening "blabla""
| 1.0 | Homogenize the way users are mentioned - Homogenize the way users are mentioned. For now, sometimes they’re mentioned by their names (no email, no company information. In notifications for example), sometimes just their first name & company (email notifications), sometimes just by their email address (guest list).
When we mention someone, wherever this is (in a notification/an invitation/a list/ an email), it should always show first name, last name, email address & company name.
The format should be: "firstName lastName (orgName) blablabla" + adding the user email on link mailto on the firstname + lastName.
e.g. [Vincent Choukroun](Vchoukroun@cascade8.com) (Cascade8) has accepted your invitation to screening "blabla""
| priority | homogenize the way users are mentioned homogenize the way users are mentioned for now sometimes they’re mentioned by their names no email no company information in notifications for example sometimes just their first name company email notifications sometimes just by their email address guest list when we mention someone wherever this is in a notification an invitation a list an email it should always show first name last name email address company name the format should be firstname lastname orgname blablabla adding the user email on link mailto on the firstname lastname e g vchoukroun com has accepted your invitation to screening blabla | 1 |
409,968 | 11,981,111,435 | IssuesEvent | 2020-04-07 10:30:22 | club-soda/club-soda-guide | https://api.github.com/repos/club-soda/club-soda-guide | opened | Remove retailer and wholesaler option on 'List a Venue' page | enhancement priority-1 | We have recently had a couple of venues who also might operate as shops try and list themselves as 'retailers' when they are listing. However, this means they are not publicly viewable and instead get pushed into the back end where retailers can be manually updated with their drinks listed by admins.
As we mainly use retailers to refer to online retailers, and monitor that as admins, we would like to remove the option for venues to list themselves as either retailers or wholesalers | 1.0 | Remove retailer and wholesaler option on 'List a Venue' page - We have recently had a couple of venues who also might operate as shops try and list themselves as 'retailers' when they are listing. However, this means they are not publicly viewable and instead get pushed into the back end where retailers can be manually updated with their drinks listed by admins.
As we mainly use retailers to refer to online retailers, and monitor that as admins, we would like to remove the option for venues to list themselves as either retailers or wholesalers | priority | remove retailer and wholesaler option on list a venue page we have recently had a couple of venues who also might operate as shops try and list themselves as retailers when they are listing however this means they are not publicly viewable and instead get pushed into the back end where retailers can be manually updated with their drinks listed by admins as we mainly use retailers to refer to online retailers and monitor that as admins we would like to remove the option for venues to list themselves as either retailers or wholesalers | 1 |
276,655 | 8,607,134,864 | IssuesEvent | 2018-11-17 19:12:16 | hassio-addons/addon-mqtt | https://api.github.com/repos/hassio-addons/addon-mqtt | closed | Connecting to CloudMQTT as a MQTT bridge | Priority: Medium Status: In progress Type: Enhancement Type: Feature | # Problem/Motivation
Looking to move over from Official MQTT server for the WebUI for debugging
--
Is there a way to configure an MQTT bridge to say cloud mqtt? I've accomplished this with the offical MQTT server using this setup: https://community.home-assistant.io/t/solved-mqtt-over-internet-aka-how-to-set-up-cloudmqtt-bridge-with-hassio-mosquitto-broker/49939/8
Does this still exist for this add on as well?
<blockquote><img src="https://community.home-assistant.io/user_avatar/community.home-assistant.io/krash/200/20614_1.png" width="48" align="right"><div><img src="https://community.home-assistant.io/uploads/default/original/1X/510301398ed5188e395eafb0983fbf8fe8ffcbcf.png" height="14"> Home Assistant Community</div><div><strong><a href="https://community.home-assistant.io/t/solved-mqtt-over-internet-aka-how-to-set-up-cloudmqtt-bridge-with-hassio-mosquitto-broker/49939/8">[solved] Mqtt over internet? aka: "How to set up cloudMQTT bridge with Hassio mosquitto broker"</a></strong></div><div>I have not worked with NodeRed at all, but i figured it out without it. Here is how i did it, by compiling info from all around the community forums: Register CloudMQTT free account. Ignore the initial un/pw and create two (in my case) new users with their own passwords. In the users page of cloudMQTT, scroll down after creating the users and click topic, select the user, add # (in case you want them in on everything) in the pattern box, then tick read/write access accordingly and add them tw...</div></blockquote> | 1.0 | Connecting to CloudMQTT as a MQTT bridge - # Problem/Motivation
Looking to move over from Official MQTT server for the WebUI for debugging
--
Is there a way to configure an MQTT bridge to say cloud mqtt? I've accomplished this with the offical MQTT server using this setup: https://community.home-assistant.io/t/solved-mqtt-over-internet-aka-how-to-set-up-cloudmqtt-bridge-with-hassio-mosquitto-broker/49939/8
Does this still exist for this add on as well?
<blockquote><img src="https://community.home-assistant.io/user_avatar/community.home-assistant.io/krash/200/20614_1.png" width="48" align="right"><div><img src="https://community.home-assistant.io/uploads/default/original/1X/510301398ed5188e395eafb0983fbf8fe8ffcbcf.png" height="14"> Home Assistant Community</div><div><strong><a href="https://community.home-assistant.io/t/solved-mqtt-over-internet-aka-how-to-set-up-cloudmqtt-bridge-with-hassio-mosquitto-broker/49939/8">[solved] Mqtt over internet? aka: "How to set up cloudMQTT bridge with Hassio mosquitto broker"</a></strong></div><div>I have not worked with NodeRed at all, but i figured it out without it. Here is how i did it, by compiling info from all around the community forums: Register CloudMQTT free account. Ignore the initial un/pw and create two (in my case) new users with their own passwords. In the users page of cloudMQTT, scroll down after creating the users and click topic, select the user, add # (in case you want them in on everything) in the pattern box, then tick read/write access accordingly and add them tw...</div></blockquote> | priority | connecting to cloudmqtt as a mqtt bridge problem motivation looking to move over from official mqtt server for the webui for debugging is there a way to configure an mqtt bridge to say cloud mqtt i ve accomplished this with the offical mqtt server using this setup does this still exist for this add on as well home assistant community i have not worked with nodered at all but i figured it out without it here is how i did it by compiling info from all around the community forums register cloudmqtt free account ignore the initial un pw and create two in my case new users with their own passwords in the users page of cloudmqtt scroll down after creating the users and click topic select the user add in case you want them in on everything in the pattern box then tick read write access accordingly and add them tw | 1 |
28,373 | 23,199,698,902 | IssuesEvent | 2022-08-01 20:05:51 | iree-org/iree | https://api.github.com/repos/iree-org/iree | opened | Better define testing tools | infrastructure | ### Issue body
Related to #9881, I'm noticing quite a few places where a CI build job needs an entire massive directory just so it can poke into it and get a single binary. Examples include getting the entire host build directory just for FileCheck in the RISCV64 build and checking out all submodules just for lit.py. We should better define the artifacts needed for testing and package them up for various builds. I think the best thing would be trying again for installable tests, but even if it's not that, we should be able to do a better job here. | 1.0 | Better define testing tools - ### Issue body
Related to #9881, I'm noticing quite a few places where a CI build job needs an entire massive directory just so it can poke into it and get a single binary. Examples include getting the entire host build directory just for FileCheck in the RISCV64 build and checking out all submodules just for lit.py. We should better define the artifacts needed for testing and package them up for various builds. I think the best thing would be trying again for installable tests, but even if it's not that, we should be able to do a better job here. | non_priority | better define testing tools issue body related to i m noticing quite a few places where a ci build job needs an entire massive directory just so it can poke into it and get a single binary examples include getting the entire host build directory just for filecheck in the build and checking out all submodules just for lit py we should better define the artifacts needed for testing and package them up for various builds i think the best thing would be trying again for installable tests but even if it s not that we should be able to do a better job here | 0 |
344,011 | 10,339,214,610 | IssuesEvent | 2019-09-03 18:47:28 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | Fix recursive exceptions from _RateLimitFilter | area: production help wanted in progress priority: high | We have a very useful piece of code, `_RateLimitFilter`, which is designed to avoid sending us a billion error emails in the event that a Zulip production server is down in a way that throws the same exception a lot. The code uses memcached to ensure we send each traceback roughly once per Zulip server per 10 minutes (or if memcached is unavailable, at most 1/process/10 minutes, since we use memcached to coordinate between processes)
There are a few problems with it:
(1) There are no tests, which makes iterating on the code a risky option.
(2) If memcached is down, there end up being some `logging.error` calls internal to the Django/memcached setup that happen inside the `cache.set()` call, and those aren't caught by the `except Exception` block around it. This ends up resulting in infinite recursion, eventually leading to `Fatal Python error: Cannot recover from stack overflow.`, since this handler is configured to run for `logging.error` in addition to `logging.exception`.
We should fix this as follows:
* Eliminate the "infinite recursion" problem by adding an outer try/finally block that sets a global (or maybe thread-local?) variable `handling_exception = True` at the start, and sets it back to false at the end.
* Add unit tests that can reproduce this error condition (I think basically we just need to `mock.patch` the `cache.get` call to either throw an exception or do a `logging.error` call for the main corner cases)
* Let's look at the actual `logging.error` message Django gives when memcached is down; I have a recollection that it has a large random ID in it, which makes the duplicate-filtering we use not work at all.
| 1.0 | Fix recursive exceptions from _RateLimitFilter - We have a very useful piece of code, `_RateLimitFilter`, which is designed to avoid sending us a billion error emails in the event that a Zulip production server is down in a way that throws the same exception a lot. The code uses memcached to ensure we send each traceback roughly once per Zulip server per 10 minutes (or if memcached is unavailable, at most 1/process/10 minutes, since we use memcached to coordinate between processes)
There are a few problems with it:
(1) There are no tests, which makes iterating on the code a risky option.
(2) If memcached is down, there end up being some `logging.error` calls internal to the Django/memcached setup that happen inside the `cache.set()` call, and those aren't caught by the `except Exception` block around it. This ends up resulting in infinite recursion, eventually leading to `Fatal Python error: Cannot recover from stack overflow.`, since this handler is configured to run for `logging.error` in addition to `logging.exception`.
We should fix this as follows:
* Eliminate the "infinite recursion" problem by adding an outer try/finally block that sets a global (or maybe thread-local?) variable `handling_exception = True` at the start, and sets it back to false at the end.
* Add unit tests that can reproduce this error condition (I think basically we just need to `mock.patch` the `cache.get` call to either throw an exception or do a `logging.error` call for the main corner cases)
* Let's look at the actual `logging.error` message Django gives when memcached is down; I have a recollection that it has a large random ID in it, which makes the duplicate-filtering we use not work at all.
| priority | fix recursive exceptions from ratelimitfilter we have a very useful piece of code ratelimitfilter which is designed to avoid sending us a billion error emails in the event that a zulip production server is down in a way that throws the same exception a lot the code uses memcached to ensure we send each traceback roughly once per zulip server per minutes or if memcached is unavailable at most process minutes since we use memcached to coordinate between processes there are a few problems with it there are no tests which makes iterating on the code a risky option if memcached is down there end up being some logging error calls internal to the django memcached setup that happen inside the cache set call and those aren t caught by the except exception block around it this ends up resulting in infinite recursion eventually leading to fatal python error cannot recover from stack overflow since this handler is configured to run for logging error in addition to logging exception we should fix this as follows eliminate the infinite recursion problem by adding an outer try finally block that sets a global or maybe thread local variable handling exception true at the start and sets it back to false at the end add unit tests that can reproduce this error condition i think basically we just need to mock patch the cache get call to either throw an exception or do a logging error call for the main corner cases let s look at the actual logging error message django gives when memcached is down i have a recollection that it has a large random id in it which makes the duplicate filtering we use not work at all | 1 |
252,710 | 8,039,317,263 | IssuesEvent | 2018-07-30 17:59:50 | idaholab/raven | https://api.github.com/repos/idaholab/raven | opened | DataObject.options missing pivotParameter explanation | improvement manuals priority_critical | --------
Issue Description
--------
It's not clear how to set the pivotParameter from the documentation.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| 1.0 | DataObject.options missing pivotParameter explanation - --------
Issue Description
--------
It's not clear how to set the pivotParameter from the documentation.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| priority | dataobject options missing pivotparameter explanation issue description it s not clear how to set the pivotparameter from the documentation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided | 1 |
729,809 | 25,145,982,771 | IssuesEvent | 2022-11-10 05:25:42 | Rehachoudhary0/hotel_testing | https://api.github.com/repos/Rehachoudhary0/hotel_testing | closed | 🐛 Bug Report: New hotel booking hotel notification. | bug app High priority | ### 👟 Reproduction steps
New hotel booking shows to hotel push notification numeric numbers (user id )only not showing the name of user.
### 👍 Expected behavior
should be show customer number also .
### 👎 Actual Behavior

.
### 🎲 App version
Version 22.10.16+01
### 💻 Operating system
Android
### 👀 Have you spent some time to check if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Code of Conduct?
- [X] I have read the [Code of Conduct](https://github.com/Rehachoudhary0/hotel_testing/blob/HEAD/CODE_OF_CONDUCT.md) | 1.0 | 🐛 Bug Report: New hotel booking hotel notification. - ### 👟 Reproduction steps
New hotel booking shows to hotel push notification numeric numbers (user id )only not showing the name of user.
### 👍 Expected behavior
should be show customer number also .
### 👎 Actual Behavior

.
### 🎲 App version
Version 22.10.16+01
### 💻 Operating system
Android
### 👀 Have you spent some time to check if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Code of Conduct?
- [X] I have read the [Code of Conduct](https://github.com/Rehachoudhary0/hotel_testing/blob/HEAD/CODE_OF_CONDUCT.md) | priority | 🐛 bug report new hotel booking hotel notification 👟 reproduction steps new hotel booking shows to hotel push notification numeric numbers user id only not showing the name of user 👍 expected behavior should be show customer number also 👎 actual behavior 🎲 app version version 💻 operating system android 👀 have you spent some time to check if this issue has been raised before i checked and didn t find similar issue 🏢 have you read the code of conduct i have read the | 1 |
258,832 | 27,582,851,705 | IssuesEvent | 2023-03-08 17:23:29 | feemstr/confluent-kafka-dotnet-721 | https://api.github.com/repos/feemstr/confluent-kafka-dotnet-721 | opened | system.net.http.4.3.0.nupkg: 1 vulnerabilities (highest severity is: 7.5) | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /src/Confluent.SchemaRegistry.Serdes.Json/Confluent.SchemaRegistry.Serdes.Json.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/feemstr/confluent-kafka-dotnet-721/commit/ff2df9cdf68ec852d156a82c43c9cae264cc54b0">ff2df9cdf68ec852d156a82c43c9cae264cc54b0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (system.net.http.4.3.0.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Direct | System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-8292</summary>
### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /src/Confluent.SchemaRegistry.Serdes.Json/Confluent.SchemaRegistry.Serdes.Json.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/feemstr/confluent-kafka-dotnet-721/commit/ff2df9cdf68ec852d156a82c43c9cae264cc54b0">ff2df9cdf68ec852d156a82c43c9cae264cc54b0</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | system.net.http.4.3.0.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /src/Confluent.SchemaRegistry.Serdes.Json/Confluent.SchemaRegistry.Serdes.Json.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/feemstr/confluent-kafka-dotnet-721/commit/ff2df9cdf68ec852d156a82c43c9cae264cc54b0">ff2df9cdf68ec852d156a82c43c9cae264cc54b0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (system.net.http.4.3.0.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Direct | System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-8292</summary>
### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /src/Confluent.SchemaRegistry.Serdes.Json/Confluent.SchemaRegistry.Serdes.Json.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/feemstr/confluent-kafka-dotnet-721/commit/ff2df9cdf68ec852d156a82c43c9cae264cc54b0">ff2df9cdf68ec852d156a82c43c9cae264cc54b0</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_priority | system net http nupkg vulnerabilities highest severity is vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file src confluent schemaregistry serdes json confluent schemaregistry serdes json csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in system net http nupkg version remediation available high system net http nupkg direct system net http microsoft powershell commands utility rc details cve vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file src confluent schemaregistry serdes json confluent schemaregistry serdes json csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy x system net http nupkg vulnerable library found in head commit a href found in base branch main vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution system net http microsoft powershell commands utility rc rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
777,513 | 27,282,991,492 | IssuesEvent | 2023-02-23 11:26:05 | o3de/o3de | https://api.github.com/repos/o3de/o3de | closed | Perform Import in Editor Settings Manager can be used without file selection | feature/editor kind/bug sig/content triage/accepted priority/minor | **Describe the bug**
It is possible to import settings in the Editor Settings Manager while no file was selected. This leads to the Manager closing docked Tools in the Editor.
This issue occurs on both Windows and Linux.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Open the Editor
2. Open Edit -> Editor Settings -> Editor Settings Manager
3. Click on the 'Perform Import' button on the bottom of the Manager
**Expected behavior**
The button is unusable without Layout file selection.
**Actual behavior**
The Manager closes every docked Tool in the Editor.
**Screenshots/Video**
2023-01-31 12-00-01.mp4
**Found in Branch**
Development
**Commit ID from [o3de/o3de](https://github.com/o3de/o3de) Repository**
adc2f3f
**Desktop/Device (please complete the following information):**
- Device: PC
- OS: Windows
- Version: 11 22H2
- CPU: I7-11700f
- GPU: Geforce RTX3070 MSI
- Memory: 32GB | 1.0 | Perform Import in Editor Settings Manager can be used without file selection - **Describe the bug**
It is possible to import settings in the Editor Settings Manager while no file was selected. This leads to the Manager closing docked Tools in the Editor.
This issue occurs on both Windows and Linux.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Open the Editor
2. Open Edit -> Editor Settings -> Editor Settings Manager
3. Click on the 'Perform Import' button on the bottom of the Manager
**Expected behavior**
The button is unusable without Layout file selection.
**Actual behavior**
The Manager closes every docked Tool in the Editor.
**Screenshots/Video**
2023-01-31 12-00-01.mp4
**Found in Branch**
Development
**Commit ID from [o3de/o3de](https://github.com/o3de/o3de) Repository**
adc2f3f
**Desktop/Device (please complete the following information):**
- Device: PC
- OS: Windows
- Version: 11 22H2
- CPU: I7-11700f
- GPU: Geforce RTX3070 MSI
- Memory: 32GB | priority | perform import in editor settings manager can be used without file selection describe the bug it is possible to import settings in the editor settings manager while no file was selected this leads to the manager closing docked tools in the editor this issue occurs on both windows and linux steps to reproduce steps to reproduce the behavior open the editor open edit editor settings editor settings manager click on the perform import button on the bottom of the manager expected behavior the button is unusable without layout file selection actual behavior the manager closes every docked tool in the editor screenshots video found in branch development commit id from repository desktop device please complete the following information device pc os windows version cpu gpu geforce msi memory | 1 |
15,628 | 11,622,014,650 | IssuesEvent | 2020-02-27 05:04:54 | GIScience/openpoiservice | https://api.github.com/repos/GIScience/openpoiservice | opened | Update dependencies to newest versions | infrastructure | Update project to work on newest packages, most notably `Flask>=1.0`, which also means `Werkzeug>=1.0`. | 1.0 | Update dependencies to newest versions - Update project to work on newest packages, most notably `Flask>=1.0`, which also means `Werkzeug>=1.0`. | non_priority | update dependencies to newest versions update project to work on newest packages most notably flask which also means werkzeug | 0 |
469,295 | 13,505,101,223 | IssuesEvent | 2020-09-13 21:09:11 | ramack/openhab2-addons | https://api.github.com/repos/ramack/openhab2-addons | closed | Reading temperature sensors connected to a DL3 | help wanted low priority | Hi,
I have successfully got the binding working with a DL3, which in turn is connected to a SKSC3HE from which I can read all values as expected.
I also have a couple of sensors connected directly to the DL3, but i can't find these in the Paper UI, i tried adding the DL3 as a device, but no luck there.

Not sure if the issue should be posted here or on the community thread, so let me know if it's the wrong place. | 1.0 | Reading temperature sensors connected to a DL3 - Hi,
I have successfully got the binding working with a DL3, which in turn is connected to a SKSC3HE from which I can read all values as expected.
I also have a couple of sensors connected directly to the DL3, but i can't find these in the Paper UI, i tried adding the DL3 as a device, but no luck there.

Not sure if the issue should be posted here or on the community thread, so let me know if it's the wrong place. | priority | reading temperature sensors connected to a hi i have successfully got the binding working with a which in turn is connected to a from which i can read all values as expected i also have a couple of sensors connected directly to the but i can t find these in the paper ui i tried adding the as a device but no luck there not sure if the issue should be posted here or on the community thread so let me know if it s the wrong place | 1 |
204,713 | 23,272,157,207 | IssuesEvent | 2022-08-05 01:09:17 | temporalio/samples-typescript | https://api.github.com/repos/temporalio/samples-typescript | closed | node-fetch-2.6.7.tgz: 1 vulnerabilities (highest severity is: 5.9) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-2.6.7.tgz</b></p></summary>
<p>A light-weight module that brings window.fetch to node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz</a></p>
<p>Path to dependency file: /activities-examples/package.json</p>
<p>Path to vulnerable library: /activities-examples/node_modules/node-fetch/package.json,/patching-api/node_modules/node-fetch/package.json,/interceptors-opentelemetry/node_modules/node-fetch/package.json,/node_modules/node-fetch/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-2596](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2596) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | node-fetch-2.6.7.tgz | Direct | 4.0.0-beta.1 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-2596</summary>
### Vulnerable Library - <b>node-fetch-2.6.7.tgz</b></p>
<p>A light-weight module that brings window.fetch to node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz</a></p>
<p>Path to dependency file: /activities-examples/package.json</p>
<p>Path to vulnerable library: /activities-examples/node_modules/node-fetch/package.json,/patching-api/node_modules/node-fetch/package.json,/interceptors-opentelemetry/node_modules/node-fetch/package.json,/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-fetch-2.6.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Denial of Service in GitHub repository node-fetch/node-fetch prior to 3.2.10.
<p>Publish Date: 2022-08-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2596>CVE-2022-2596</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-2596">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-2596</a></p>
<p>Release Date: 2022-08-01</p>
<p>Fix Resolution: 4.0.0-beta.1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | node-fetch-2.6.7.tgz: 1 vulnerabilities (highest severity is: 5.9) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-2.6.7.tgz</b></p></summary>
<p>A light-weight module that brings window.fetch to node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz</a></p>
<p>Path to dependency file: /activities-examples/package.json</p>
<p>Path to vulnerable library: /activities-examples/node_modules/node-fetch/package.json,/patching-api/node_modules/node-fetch/package.json,/interceptors-opentelemetry/node_modules/node-fetch/package.json,/node_modules/node-fetch/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-2596](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2596) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | node-fetch-2.6.7.tgz | Direct | 4.0.0-beta.1 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-2596</summary>
### Vulnerable Library - <b>node-fetch-2.6.7.tgz</b></p>
<p>A light-weight module that brings window.fetch to node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz</a></p>
<p>Path to dependency file: /activities-examples/package.json</p>
<p>Path to vulnerable library: /activities-examples/node_modules/node-fetch/package.json,/patching-api/node_modules/node-fetch/package.json,/interceptors-opentelemetry/node_modules/node-fetch/package.json,/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-fetch-2.6.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Denial of Service in GitHub repository node-fetch/node-fetch prior to 3.2.10.
<p>Publish Date: 2022-08-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2596>CVE-2022-2596</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-2596">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-2596</a></p>
<p>Release Date: 2022-08-01</p>
<p>Fix Resolution: 4.0.0-beta.1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_priority | node fetch tgz vulnerabilities highest severity is autoclosed vulnerable library node fetch tgz a light weight module that brings window fetch to node js library home page a href path to dependency file activities examples package json path to vulnerable library activities examples node modules node fetch package json patching api node modules node fetch package json interceptors opentelemetry node modules node fetch package json node modules node fetch package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium node fetch tgz direct beta details cve vulnerable library node fetch tgz a light weight module that brings window fetch to node js library home page a href path to dependency file activities examples package json path to vulnerable library activities examples node modules node fetch package json patching api node modules node fetch package json interceptors opentelemetry node modules node fetch package json node modules node fetch package json dependency hierarchy x node fetch tgz vulnerable library found in head commit a href found in base branch main vulnerability details denial of service in github repository node fetch node fetch prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution beta rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
586,335 | 17,575,475,114 | IssuesEvent | 2021-08-15 14:22:43 | DSC-University-of-Seoul/2021-spring-project | https://api.github.com/repos/DSC-University-of-Seoul/2021-spring-project | closed | [스프린트 #9] 로그인 기능 구현 | Priority ❗️ Medium Type 🛠 FEATURE Issue ⏩ IN PROGRESS | # 기능 Issue
Linked Feature backlog : #KDSK-107
## 필요한 기능
백엔드와 프론트엔드간 로그인 기능 연계
## 설명
- 프론트엔드에서 로그인 기능 구현을 진행중입니다.
- 백엔드에 POST로 입력받은 유저 정보를 전송할 계획입니다.
- 백엔드에서 로그인 성공 / 실패 여부가 어떤 형식으로 반환되는지 명시 부탁드립니다.
- 추가로 POST 시 필요한 헤더 정보가 있으면 명시 부탁드립니다.
## 담당자 할당 확인
담당자: @rxdcxdrnine
- [ ] 담당자 접수 확인
## 우선순위 확인
- [ ] 담당자의 우선순위 확인
- [x] PM의 최종 우선순위 확인
###
| 1.0 | [스프린트 #9] 로그인 기능 구현 - # 기능 Issue
Linked Feature backlog : #KDSK-107
## 필요한 기능
백엔드와 프론트엔드간 로그인 기능 연계
## 설명
- 프론트엔드에서 로그인 기능 구현을 진행중입니다.
- 백엔드에 POST로 입력받은 유저 정보를 전송할 계획입니다.
- 백엔드에서 로그인 성공 / 실패 여부가 어떤 형식으로 반환되는지 명시 부탁드립니다.
- 추가로 POST 시 필요한 헤더 정보가 있으면 명시 부탁드립니다.
## 담당자 할당 확인
담당자: @rxdcxdrnine
- [ ] 담당자 접수 확인
## 우선순위 확인
- [ ] 담당자의 우선순위 확인
- [x] PM의 최종 우선순위 확인
###
| priority | 로그인 기능 구현 기능 issue linked feature backlog kdsk 필요한 기능 백엔드와 프론트엔드간 로그인 기능 연계 설명 프론트엔드에서 로그인 기능 구현을 진행중입니다 백엔드에 post로 입력받은 유저 정보를 전송할 계획입니다 백엔드에서 로그인 성공 실패 여부가 어떤 형식으로 반환되는지 명시 부탁드립니다 추가로 post 시 필요한 헤더 정보가 있으면 명시 부탁드립니다 담당자 할당 확인 담당자 rxdcxdrnine 담당자 접수 확인 우선순위 확인 담당자의 우선순위 확인 pm의 최종 우선순위 확인 | 1 |
448,641 | 12,955,030,692 | IssuesEvent | 2020-07-20 05:22:11 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | Edit service, labelSelector is not set, error occur after saving | area/console kind/bug kind/need-to-verify priority/medium | **Describe the Bug**
The edit service page should not be saved successfully if LabelSelector is not set
<img width="865" alt="selector" src="https://user-images.githubusercontent.com/36271543/87624649-d30c3c80-c75a-11ea-9555-55c1b78514c0.png">
If labelSelector was not set, it would receive an error: namespaces "undefined" not found
<img width="903" alt="error5" src="https://user-images.githubusercontent.com/36271543/87624776-2088a980-c75b-11ea-833f-553c7534cbd1.png">
**Versions Used**
KubeSphere:3.0.0
Kubernetes:
host-v1.16.12
member1-v1.18.5
member2-v1.17.8
**Environment**
host: 1node /ubuntu 16.04 4cpu/16g
member1: 2 nodes /ubuntu 16.04 4cpu/16g
member1: 2 nodes /centos7 8cpu/16g
/kind bug
/area console
/assign @leoendless
/milestone 3.0.0
/priority medium
| 1.0 | Edit service, labelSelector is not set, error occur after saving - **Describe the Bug**
The edit service page should not be saved successfully if LabelSelector is not set
<img width="865" alt="selector" src="https://user-images.githubusercontent.com/36271543/87624649-d30c3c80-c75a-11ea-9555-55c1b78514c0.png">
If labelSelector was not set, it would receive an error: namespaces "undefined" not found
<img width="903" alt="error5" src="https://user-images.githubusercontent.com/36271543/87624776-2088a980-c75b-11ea-833f-553c7534cbd1.png">
**Versions Used**
KubeSphere:3.0.0
Kubernetes:
host-v1.16.12
member1-v1.18.5
member2-v1.17.8
**Environment**
host: 1node /ubuntu 16.04 4cpu/16g
member1: 2 nodes /ubuntu 16.04 4cpu/16g
member1: 2 nodes /centos7 8cpu/16g
/kind bug
/area console
/assign @leoendless
/milestone 3.0.0
/priority medium
| priority | edit service labelselector is not set error occur after saving describe the bug the edit service page should not be saved successfully if labelselector is not set img width alt selector src if labelselector was not set it would receive an error namespaces undefined not found img width alt src versions used kubesphere kubernetes host environment host ubuntu nodes ubuntu nodes kind bug area console assign leoendless milestone priority medium | 1 |
604,638 | 18,716,006,964 | IssuesEvent | 2021-11-03 04:52:51 | AY2122S1-CS2103T-T15-2/tp | https://api.github.com/repos/AY2122S1-CS2103T-T15-2/tp | closed | A help/tutorial/guide. | type.Story priority.Medium type.Task | As a user, I want to know how to use the application, so that I can troubleshoot any problems myself easily. | 1.0 | A help/tutorial/guide. - As a user, I want to know how to use the application, so that I can troubleshoot any problems myself easily. | priority | a help tutorial guide as a user i want to know how to use the application so that i can troubleshoot any problems myself easily | 1 |
64,364 | 14,662,054,601 | IssuesEvent | 2020-12-29 06:05:33 | tamirverthim/NodeGoat | https://api.github.com/repos/tamirverthim/NodeGoat | opened | CVE-2019-1010266 (Medium) detected in lodash-4.13.1.tgz, lodash-2.4.2.tgz | security vulnerability | ## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.13.1.tgz</b>, <b>lodash-2.4.2.tgz</b></p></summary>
<p>
<details><summary><b>lodash-4.13.1.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz">https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/nyc/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- nyc-7.1.0.tgz
- istanbul-lib-instrument-1.1.0-alpha.4.tgz
- babel-generator-6.11.4.tgz
- :x: **lodash-4.13.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/zaproxy/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- zaproxy-0.2.0.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/3de6c5862c1fef83d38a1fec17b579f1a5e328fb">3de6c5862c1fef83d38a1fec17b579f1a5e328fb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2019-07-17</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.13.1","isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-lib-instrument:1.1.0-alpha.4;babel-generator:6.11.4;lodash:4.13.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.2","isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;lodash:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2019-1010266","vulnerabilityDetails":"lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-1010266 (Medium) detected in lodash-4.13.1.tgz, lodash-2.4.2.tgz - ## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.13.1.tgz</b>, <b>lodash-2.4.2.tgz</b></p></summary>
<p>
<details><summary><b>lodash-4.13.1.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz">https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/nyc/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- nyc-7.1.0.tgz
- istanbul-lib-instrument-1.1.0-alpha.4.tgz
- babel-generator-6.11.4.tgz
- :x: **lodash-4.13.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/zaproxy/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- zaproxy-0.2.0.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/3de6c5862c1fef83d38a1fec17b579f1a5e328fb">3de6c5862c1fef83d38a1fec17b579f1a5e328fb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2019-07-17</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.13.1","isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-lib-instrument:1.1.0-alpha.4;babel-generator:6.11.4;lodash:4.13.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.2","isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;lodash:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2019-1010266","vulnerabilityDetails":"lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in lodash tgz lodash tgz cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash modular utilities library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules nyc node modules lodash package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz istanbul lib instrument alpha tgz babel generator tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules zaproxy node modules lodash package json dependency hierarchy zaproxy tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is vulnerabilityurl | 0 |
198,170 | 14,966,708,446 | IssuesEvent | 2021-01-27 14:53:35 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | closed | Flake: TestDiagnose/bazel times out | meta/test-flake | [TestDiagnose/bazel](https://api.travis-ci.com/v3/job/473483182/log.txt) failure. What's odd here is that from the output ("Ran in 2 minutes ...") the test seems to have succeeded, but goroutine 270 seems to indicate that it's still executing an external command.
Is it possible this test is running in parallel? (though wouldn't there be a log statement for the command?)
```
=== RUN TestDiagnose/bazel
time="2021-01-20T17:35:47Z" level=info msg="[skaffold diagnose]"
time="2021-01-20T17:35:47Z" level=info msg="Ran in 2 minutes 0.245 second"
panic: test timed out after 20m0s
goroutine 240 [running]:
testing.(*M).startAlarm.func1()
/usr/local/go/src/testing/testing.go:1618 +0xe5
created by time.goFunc
/usr/local/go/src/time/sleep.go:167 +0x45
goroutine 1 [chan receive]:
testing.(*T).Run(0xc0008b9200, 0x214e68a, 0xc, 0x22ef668, 0x49b801)
/usr/local/go/src/testing/testing.go:1169 +0x2da
testing.runTests.func1(0xc000109080)
/usr/local/go/src/testing/testing.go:1439 +0x78
testing.tRunner(0xc000109080, 0xc0008ffde0)
/usr/local/go/src/testing/testing.go:1123 +0xef
testing.runTests(0xc00031c1e0, 0x33628a0, 0x46, 0x46, 0xbffa385d196693d2, 0x11767f489d8, 0x337a580, 0x40f6b0)
/usr/local/go/src/testing/testing.go:1437 +0x2fe
testing.(*M).Run(0xc00070c000, 0x0)
/usr/local/go/src/testing/testing.go:1345 +0x1eb
main.main()
_testmain.go:181 +0x138
goroutine 19 [chan receive]:
k8s.io/klog/v2.(*loggingT).flushDaemon(0x337ad60)
/root/go/pkg/mod/k8s.io/klog/v2@v2.2.0/klog.go:1131 +0x8b
created by k8s.io/klog/v2.init.0
/root/go/pkg/mod/k8s.io/klog/v2@v2.2.0/klog.go:416 +0xd8
goroutine 6 [chan receive]:
github.com/golang/glog.(*loggingT).flushDaemon(0x337abc0)
/root/go/pkg/mod/github.com/golang/glog@v0.0.0-20160126235308-23def4e6c14b/glog.go:882 +0x8b
created by github.com/golang/glog.init.0
/root/go/pkg/mod/github.com/golang/glog@v0.0.0-20160126235308-23def4e6c14b/glog.go:410 +0x274
goroutine 7 [select]:
go.opencensus.io/stats/view.(*worker).start(0xc00070c700)
/root/go/pkg/mod/go.opencensus.io@v0.22.5/stats/view/worker.go:276 +0x105
created by go.opencensus.io/stats/view.init.0
/root/go/pkg/mod/go.opencensus.io@v0.22.5/stats/view/worker.go:34 +0x68
goroutine 213 [IO wait]:
internal/poll.runtime_pollWait(0x7f8e3ce36c30, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc000930918, 0x72, 0x24a1800, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc000930900, 0xc000a80000, 0x6986, 0x6986, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
net.(*netFD).Read(0xc000930900, 0xc000a80000, 0x6986, 0x6986, 0x203000, 0x7731bb, 0xc0003b64e0)
/usr/local/go/src/net/fd_posix.go:55 +0x4f
net.(*conn).Read(0xc000010058, 0xc000a80000, 0x6986, 0x6986, 0x0, 0x0, 0x0)
/usr/local/go/src/net/net.go:182 +0x8e
crypto/tls.(*atLeastReader).Read(0xc0002de140, 0xc000a80000, 0x6986, 0x6986, 0x267, 0x6946, 0xc000079710)
/usr/local/go/src/crypto/tls/conn.go:779 +0x62
bytes.(*Buffer).ReadFrom(0xc0003b6600, 0x2499260, 0xc0002de140, 0x40d3c5, 0x1e495a0, 0x20b94e0)
/usr/local/go/src/bytes/buffer.go:204 +0xb1
crypto/tls.(*Conn).readFromUntil(0xc0003b6380, 0x249ef20, 0xc000010058, 0x5, 0xc000010058, 0x256)
/usr/local/go/src/crypto/tls/conn.go:801 +0xf3
crypto/tls.(*Conn).readRecordOrCCS(0xc0003b6380, 0x0, 0x0, 0xc000079d18)
/usr/local/go/src/crypto/tls/conn.go:608 +0x115
crypto/tls.(*Conn).readRecord(...)
/usr/local/go/src/crypto/tls/conn.go:576
crypto/tls.(*Conn).Read(0xc0003b6380, 0xc0004df000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/crypto/tls/conn.go:1252 +0x15f
bufio.(*Reader).Read(0xc00009bbc0, 0xc000116498, 0x9, 0x9, 0xc000079d18, 0x22f3e00, 0x97096b)
/usr/local/go/src/bufio/bufio.go:227 +0x222
io.ReadAtLeast(0x2499000, 0xc00009bbc0, 0xc000116498, 0x9, 0x9, 0x9, 0xc000114030, 0x0, 0x24994c0)
/usr/local/go/src/io/io.go:314 +0x87
io.ReadFull(...)
/usr/local/go/src/io/io.go:333
golang.org/x/net/http2.readFrameHeader(0xc000116498, 0x9, 0x9, 0x2499000, 0xc00009bbc0, 0x0, 0x0, 0xc0004360c0, 0x0)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/frame.go:237 +0x89
golang.org/x/net/http2.(*Framer).ReadFrame(0xc000116460, 0xc0004360c0, 0x0, 0x0, 0x0)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/frame.go:492 +0xa5
golang.org/x/net/http2.(*clientConnReadLoop).run(0xc000079fa8, 0x0, 0x0)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/transport.go:1819 +0xd8
golang.org/x/net/http2.(*ClientConn).readLoop(0xc000582f00)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/transport.go:1741 +0x6f
created by golang.org/x/net/http2.(*Transport).newClientConn
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/transport.go:705 +0x6c5
goroutine 9 [chan receive, 20 minutes]:
github.com/GoogleContainerTools/skaffold/pkg/skaffold/event.newHandler.func1(0xc0000e2580)
/skaffold/pkg/skaffold/event/event.go:59 +0x45
created by github.com/GoogleContainerTools/skaffold/pkg/skaffold/event.newHandler
/skaffold/pkg/skaffold/event/event.go:57 +0x90
goroutine 10 [chan receive, 20 minutes]:
github.com/rjeczalik/notify.(*nonrecursiveTree).dispatch(0xc00009b680, 0xc00009b500)
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:36 +0xb6
created by github.com/rjeczalik/notify.newNonrecursiveTree
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:29 +0xe5
goroutine 11 [chan receive, 20 minutes]:
github.com/rjeczalik/notify.(*nonrecursiveTree).internal(0xc00009b680, 0xc00009b5c0)
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:81 +0x58
created by github.com/rjeczalik/notify.newNonrecursiveTree
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:30 +0x111
goroutine 269 [chan receive]:
testing.(*T).Run(0xc0008b9380, 0xc0009cc7d9, 0x5, 0xc000a5e8d0, 0x20)
/usr/local/go/src/testing/testing.go:1169 +0x2da
github.com/GoogleContainerTools/skaffold/integration.TestDiagnose(0xc0008b9200)
/skaffold/integration/diagnose_test.go:38 +0x172
testing.tRunner(0xc0008b9200, 0x22ef668)
/usr/local/go/src/testing/testing.go:1123 +0xef
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:1168 +0x2b3
goroutine 221 [chan send, 4 minutes]:
k8s.io/apimachinery/pkg/watch.(*StreamWatcher).receive(0xc000b3f240)
/root/go/pkg/mod/k8s.io/apimachinery@v0.19.4/pkg/watch/streamwatcher.go:127 +0x12d
created by k8s.io/apimachinery/pkg/watch.NewStreamWatcher
/root/go/pkg/mod/k8s.io/apimachinery@v0.19.4/pkg/watch/streamwatcher.go:71 +0xbe
goroutine 271 [IO wait]:
internal/poll.runtime_pollWait(0x7f8e3ce36978, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc000334d98, 0x72, 0x24a1801, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc000334d80, 0xc00066929b, 0x565, 0x565, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
os.(*File).read(...)
/usr/local/go/src/os/file_posix.go:31
os.(*File).Read(0xc00042cb08, 0xc00066929b, 0x565, 0x565, 0x9b, 0x0, 0x0)
/usr/local/go/src/os/file.go:116 +0x71
bytes.(*Buffer).ReadFrom(0xc000660660, 0x249f320, 0xc00042cb08, 0x7f8e14305a58, 0xc000660660, 0x1)
/usr/local/go/src/bytes/buffer.go:204 +0xb1
io.copyBuffer(0x2499060, 0xc000660660, 0x249f320, 0xc00042cb08, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc0000f57c0)
/usr/local/go/src/io/io.go:395 +0x2ff
io.Copy(...)
/usr/local/go/src/io/io.go:368
os/exec.(*Cmd).writerDescriptor.func1(0xc0008b9380, 0x22f3fe8)
/usr/local/go/src/os/exec/exec.go:311 +0x65
os/exec.(*Cmd).Start.func1(0xc000b60160, 0xc000475360)
/usr/local/go/src/os/exec/exec.go:441 +0x27
created by os/exec.(*Cmd).Start
/usr/local/go/src/os/exec/exec.go:440 +0x629
goroutine 272 [IO wait]:
internal/poll.runtime_pollWait(0x7f8e3ce36b48, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc000334f18, 0x72, 0x24a1801, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc000334f00, 0xc000682000, 0x8000, 0x8000, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
os.(*File).read(...)
/usr/local/go/src/os/file_posix.go:31
os.(*File).Read(0xc00042cb20, 0xc000682000, 0x8000, 0x8000, 0x0, 0x0, 0x0)
/usr/local/go/src/os/file.go:116 +0x71
io.copyBuffer(0x249f440, 0xc00064ecd0, 0x249f320, 0xc00042cb20, 0xc000682000, 0x8000, 0x8000, 0x0, 0x0, 0x0)
/usr/local/go/src/io/io.go:409 +0x12c
io.Copy(...)
/usr/local/go/src/io/io.go:368
os/exec.(*Cmd).writerDescriptor.func1(0x0, 0x0)
/usr/local/go/src/os/exec/exec.go:311 +0x65
os/exec.(*Cmd).Start.func1(0xc000b60160, 0xc000475420)
/usr/local/go/src/os/exec/exec.go:441 +0x27
created by os/exec.(*Cmd).Start
/usr/local/go/src/os/exec/exec.go:440 +0x629
goroutine 273 [select]:
os/exec.(*Cmd).Start.func2(0xc000b60160)
/usr/local/go/src/os/exec/exec.go:449 +0xd5
created by os/exec.(*Cmd).Start
/usr/local/go/src/os/exec/exec.go:448 +0x6c5
goroutine 270 [syscall]:
syscall.Syscall6(0xf7, 0x1, 0x12da, 0xc0006bbad0, 0x1000004, 0x0, 0x0, 0x4b1201, 0xc000334fc0, 0xc0006bbb10)
/usr/local/go/src/syscall/asm_linux_amd64.s:41 +0x5
os.(*Process).blockUntilWaitable(0xc000510690, 0x203000, 0x0, 0x0)
/usr/local/go/src/os/wait_waitid.go:32 +0xa6
os.(*Process).wait(0xc000510690, 0x22f3880, 0x22f3888, 0x22f3878)
/usr/local/go/src/os/exec_unix.go:22 +0x39
os.(*Process).Wait(...)
/usr/local/go/src/os/exec.go:125
os/exec.(*Cmd).Wait(0xc000b60160, 0x0, 0x0)
/usr/local/go/src/os/exec/exec.go:507 +0x65
os/exec.(*Cmd).Run(0xc000b60160, 0xc00064ecd0, 0x2fb71770005ac3a5)
/usr/local/go/src/os/exec/exec.go:341 +0x5c
os/exec.(*Cmd).Output(0xc000b60160, 0x11742aa4353, 0x337a580, 0x1, 0x1, 0x106)
/usr/local/go/src/os/exec/exec.go:546 +0x92
github.com/GoogleContainerTools/skaffold/integration/skaffold.(*RunBuilder).RunOrFailOutput(0xc0006bbec0, 0xc0008b9380, 0x0, 0x0, 0x0)
/skaffold/integration/skaffold/helper.go:275 +0x1a9
github.com/GoogleContainerTools/skaffold/integration/skaffold.(*RunBuilder).RunOrFail(...)
/skaffold/integration/skaffold/helper.go:229
github.com/GoogleContainerTools/skaffold/integration.TestDiagnose.func1(0xc0008b9380)
/skaffold/integration/diagnose_test.go:45 +0x1e5
testing.tRunner(0xc0008b9380, 0xc000a5e8d0)
/usr/local/go/src/testing/testing.go:1123 +0xef
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:1168 +0x2b3
goroutine 71 [IO wait, 19 minutes]:
internal/poll.runtime_pollWait(0x7f8e3ce36d18, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc0007d6718, 0x72, 0x24a1800, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc0007d6700, 0xc000800000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
net.(*netFD).Read(0xc0007d6700, 0xc000800000, 0x1000, 0x1000, 0x43c5dc, 0xc00007cb58, 0x469680)
/usr/local/go/src/net/fd_posix.go:55 +0x4f
net.(*conn).Read(0xc000122e88, 0xc000800000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/net/net.go:182 +0x8e
net/http.(*persistConn).Read(0xc0007f06c0, 0xc000800000, 0x1000, 0x1000, 0xc0007f4240, 0xc00007cc58, 0x407555)
/usr/local/go/src/net/http/transport.go:1887 +0x77
bufio.(*Reader).fill(0xc00075f320)
/usr/local/go/src/bufio/bufio.go:101 +0x105
bufio.(*Reader).Peek(0xc00075f320, 0x1, 0x0, 0x0, 0x1, 0x0, 0xc000745320)
/usr/local/go/src/bufio/bufio.go:139 +0x4f
net/http.(*persistConn).readLoop(0xc0007f06c0)
/usr/local/go/src/net/http/transport.go:2040 +0x1a8
created by net/http.(*Transport).dialConn
/usr/local/go/src/net/http/transport.go:1708 +0xcb7
goroutine 72 [select, 19 minutes]:
net/http.(*persistConn).writeLoop(0xc0007f06c0)
/usr/local/go/src/net/http/transport.go:2340 +0x11c
created by net/http.(*Transport).dialConn
/usr/local/go/src/net/http/transport.go:1709 +0xcdc
FAIL github.com/GoogleContainerTools/skaffold/integration 1200.071s
``` | 1.0 | Flake: TestDiagnose/bazel times out - [TestDiagnose/bazel](https://api.travis-ci.com/v3/job/473483182/log.txt) failure. What's odd here is that from the output ("Ran in 2 minutes ...") the test seems to have succeeded, but goroutine 270 seems to indicate that it's still executing an external command.
Is it possible this test is running in parallel? (though wouldn't there be a log statement for the command?)
```
=== RUN TestDiagnose/bazel
time="2021-01-20T17:35:47Z" level=info msg="[skaffold diagnose]"
time="2021-01-20T17:35:47Z" level=info msg="Ran in 2 minutes 0.245 second"
panic: test timed out after 20m0s
goroutine 240 [running]:
testing.(*M).startAlarm.func1()
/usr/local/go/src/testing/testing.go:1618 +0xe5
created by time.goFunc
/usr/local/go/src/time/sleep.go:167 +0x45
goroutine 1 [chan receive]:
testing.(*T).Run(0xc0008b9200, 0x214e68a, 0xc, 0x22ef668, 0x49b801)
/usr/local/go/src/testing/testing.go:1169 +0x2da
testing.runTests.func1(0xc000109080)
/usr/local/go/src/testing/testing.go:1439 +0x78
testing.tRunner(0xc000109080, 0xc0008ffde0)
/usr/local/go/src/testing/testing.go:1123 +0xef
testing.runTests(0xc00031c1e0, 0x33628a0, 0x46, 0x46, 0xbffa385d196693d2, 0x11767f489d8, 0x337a580, 0x40f6b0)
/usr/local/go/src/testing/testing.go:1437 +0x2fe
testing.(*M).Run(0xc00070c000, 0x0)
/usr/local/go/src/testing/testing.go:1345 +0x1eb
main.main()
_testmain.go:181 +0x138
goroutine 19 [chan receive]:
k8s.io/klog/v2.(*loggingT).flushDaemon(0x337ad60)
/root/go/pkg/mod/k8s.io/klog/v2@v2.2.0/klog.go:1131 +0x8b
created by k8s.io/klog/v2.init.0
/root/go/pkg/mod/k8s.io/klog/v2@v2.2.0/klog.go:416 +0xd8
goroutine 6 [chan receive]:
github.com/golang/glog.(*loggingT).flushDaemon(0x337abc0)
/root/go/pkg/mod/github.com/golang/glog@v0.0.0-20160126235308-23def4e6c14b/glog.go:882 +0x8b
created by github.com/golang/glog.init.0
/root/go/pkg/mod/github.com/golang/glog@v0.0.0-20160126235308-23def4e6c14b/glog.go:410 +0x274
goroutine 7 [select]:
go.opencensus.io/stats/view.(*worker).start(0xc00070c700)
/root/go/pkg/mod/go.opencensus.io@v0.22.5/stats/view/worker.go:276 +0x105
created by go.opencensus.io/stats/view.init.0
/root/go/pkg/mod/go.opencensus.io@v0.22.5/stats/view/worker.go:34 +0x68
goroutine 213 [IO wait]:
internal/poll.runtime_pollWait(0x7f8e3ce36c30, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc000930918, 0x72, 0x24a1800, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc000930900, 0xc000a80000, 0x6986, 0x6986, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
net.(*netFD).Read(0xc000930900, 0xc000a80000, 0x6986, 0x6986, 0x203000, 0x7731bb, 0xc0003b64e0)
/usr/local/go/src/net/fd_posix.go:55 +0x4f
net.(*conn).Read(0xc000010058, 0xc000a80000, 0x6986, 0x6986, 0x0, 0x0, 0x0)
/usr/local/go/src/net/net.go:182 +0x8e
crypto/tls.(*atLeastReader).Read(0xc0002de140, 0xc000a80000, 0x6986, 0x6986, 0x267, 0x6946, 0xc000079710)
/usr/local/go/src/crypto/tls/conn.go:779 +0x62
bytes.(*Buffer).ReadFrom(0xc0003b6600, 0x2499260, 0xc0002de140, 0x40d3c5, 0x1e495a0, 0x20b94e0)
/usr/local/go/src/bytes/buffer.go:204 +0xb1
crypto/tls.(*Conn).readFromUntil(0xc0003b6380, 0x249ef20, 0xc000010058, 0x5, 0xc000010058, 0x256)
/usr/local/go/src/crypto/tls/conn.go:801 +0xf3
crypto/tls.(*Conn).readRecordOrCCS(0xc0003b6380, 0x0, 0x0, 0xc000079d18)
/usr/local/go/src/crypto/tls/conn.go:608 +0x115
crypto/tls.(*Conn).readRecord(...)
/usr/local/go/src/crypto/tls/conn.go:576
crypto/tls.(*Conn).Read(0xc0003b6380, 0xc0004df000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/crypto/tls/conn.go:1252 +0x15f
bufio.(*Reader).Read(0xc00009bbc0, 0xc000116498, 0x9, 0x9, 0xc000079d18, 0x22f3e00, 0x97096b)
/usr/local/go/src/bufio/bufio.go:227 +0x222
io.ReadAtLeast(0x2499000, 0xc00009bbc0, 0xc000116498, 0x9, 0x9, 0x9, 0xc000114030, 0x0, 0x24994c0)
/usr/local/go/src/io/io.go:314 +0x87
io.ReadFull(...)
/usr/local/go/src/io/io.go:333
golang.org/x/net/http2.readFrameHeader(0xc000116498, 0x9, 0x9, 0x2499000, 0xc00009bbc0, 0x0, 0x0, 0xc0004360c0, 0x0)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/frame.go:237 +0x89
golang.org/x/net/http2.(*Framer).ReadFrame(0xc000116460, 0xc0004360c0, 0x0, 0x0, 0x0)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/frame.go:492 +0xa5
golang.org/x/net/http2.(*clientConnReadLoop).run(0xc000079fa8, 0x0, 0x0)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/transport.go:1819 +0xd8
golang.org/x/net/http2.(*ClientConn).readLoop(0xc000582f00)
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/transport.go:1741 +0x6f
created by golang.org/x/net/http2.(*Transport).newClientConn
/root/go/pkg/mod/golang.org/x/net@v0.0.0-20201202161906-c7110b5ffcbb/http2/transport.go:705 +0x6c5
goroutine 9 [chan receive, 20 minutes]:
github.com/GoogleContainerTools/skaffold/pkg/skaffold/event.newHandler.func1(0xc0000e2580)
/skaffold/pkg/skaffold/event/event.go:59 +0x45
created by github.com/GoogleContainerTools/skaffold/pkg/skaffold/event.newHandler
/skaffold/pkg/skaffold/event/event.go:57 +0x90
goroutine 10 [chan receive, 20 minutes]:
github.com/rjeczalik/notify.(*nonrecursiveTree).dispatch(0xc00009b680, 0xc00009b500)
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:36 +0xb6
created by github.com/rjeczalik/notify.newNonrecursiveTree
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:29 +0xe5
goroutine 11 [chan receive, 20 minutes]:
github.com/rjeczalik/notify.(*nonrecursiveTree).internal(0xc00009b680, 0xc00009b5c0)
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:81 +0x58
created by github.com/rjeczalik/notify.newNonrecursiveTree
/root/go/pkg/mod/github.com/rjeczalik/notify@v0.9.2/tree_nonrecursive.go:30 +0x111
goroutine 269 [chan receive]:
testing.(*T).Run(0xc0008b9380, 0xc0009cc7d9, 0x5, 0xc000a5e8d0, 0x20)
/usr/local/go/src/testing/testing.go:1169 +0x2da
github.com/GoogleContainerTools/skaffold/integration.TestDiagnose(0xc0008b9200)
/skaffold/integration/diagnose_test.go:38 +0x172
testing.tRunner(0xc0008b9200, 0x22ef668)
/usr/local/go/src/testing/testing.go:1123 +0xef
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:1168 +0x2b3
goroutine 221 [chan send, 4 minutes]:
k8s.io/apimachinery/pkg/watch.(*StreamWatcher).receive(0xc000b3f240)
/root/go/pkg/mod/k8s.io/apimachinery@v0.19.4/pkg/watch/streamwatcher.go:127 +0x12d
created by k8s.io/apimachinery/pkg/watch.NewStreamWatcher
/root/go/pkg/mod/k8s.io/apimachinery@v0.19.4/pkg/watch/streamwatcher.go:71 +0xbe
goroutine 271 [IO wait]:
internal/poll.runtime_pollWait(0x7f8e3ce36978, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc000334d98, 0x72, 0x24a1801, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc000334d80, 0xc00066929b, 0x565, 0x565, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
os.(*File).read(...)
/usr/local/go/src/os/file_posix.go:31
os.(*File).Read(0xc00042cb08, 0xc00066929b, 0x565, 0x565, 0x9b, 0x0, 0x0)
/usr/local/go/src/os/file.go:116 +0x71
bytes.(*Buffer).ReadFrom(0xc000660660, 0x249f320, 0xc00042cb08, 0x7f8e14305a58, 0xc000660660, 0x1)
/usr/local/go/src/bytes/buffer.go:204 +0xb1
io.copyBuffer(0x2499060, 0xc000660660, 0x249f320, 0xc00042cb08, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc0000f57c0)
/usr/local/go/src/io/io.go:395 +0x2ff
io.Copy(...)
/usr/local/go/src/io/io.go:368
os/exec.(*Cmd).writerDescriptor.func1(0xc0008b9380, 0x22f3fe8)
/usr/local/go/src/os/exec/exec.go:311 +0x65
os/exec.(*Cmd).Start.func1(0xc000b60160, 0xc000475360)
/usr/local/go/src/os/exec/exec.go:441 +0x27
created by os/exec.(*Cmd).Start
/usr/local/go/src/os/exec/exec.go:440 +0x629
goroutine 272 [IO wait]:
internal/poll.runtime_pollWait(0x7f8e3ce36b48, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc000334f18, 0x72, 0x24a1801, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc000334f00, 0xc000682000, 0x8000, 0x8000, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
os.(*File).read(...)
/usr/local/go/src/os/file_posix.go:31
os.(*File).Read(0xc00042cb20, 0xc000682000, 0x8000, 0x8000, 0x0, 0x0, 0x0)
/usr/local/go/src/os/file.go:116 +0x71
io.copyBuffer(0x249f440, 0xc00064ecd0, 0x249f320, 0xc00042cb20, 0xc000682000, 0x8000, 0x8000, 0x0, 0x0, 0x0)
/usr/local/go/src/io/io.go:409 +0x12c
io.Copy(...)
/usr/local/go/src/io/io.go:368
os/exec.(*Cmd).writerDescriptor.func1(0x0, 0x0)
/usr/local/go/src/os/exec/exec.go:311 +0x65
os/exec.(*Cmd).Start.func1(0xc000b60160, 0xc000475420)
/usr/local/go/src/os/exec/exec.go:441 +0x27
created by os/exec.(*Cmd).Start
/usr/local/go/src/os/exec/exec.go:440 +0x629
goroutine 273 [select]:
os/exec.(*Cmd).Start.func2(0xc000b60160)
/usr/local/go/src/os/exec/exec.go:449 +0xd5
created by os/exec.(*Cmd).Start
/usr/local/go/src/os/exec/exec.go:448 +0x6c5
goroutine 270 [syscall]:
syscall.Syscall6(0xf7, 0x1, 0x12da, 0xc0006bbad0, 0x1000004, 0x0, 0x0, 0x4b1201, 0xc000334fc0, 0xc0006bbb10)
/usr/local/go/src/syscall/asm_linux_amd64.s:41 +0x5
os.(*Process).blockUntilWaitable(0xc000510690, 0x203000, 0x0, 0x0)
/usr/local/go/src/os/wait_waitid.go:32 +0xa6
os.(*Process).wait(0xc000510690, 0x22f3880, 0x22f3888, 0x22f3878)
/usr/local/go/src/os/exec_unix.go:22 +0x39
os.(*Process).Wait(...)
/usr/local/go/src/os/exec.go:125
os/exec.(*Cmd).Wait(0xc000b60160, 0x0, 0x0)
/usr/local/go/src/os/exec/exec.go:507 +0x65
os/exec.(*Cmd).Run(0xc000b60160, 0xc00064ecd0, 0x2fb71770005ac3a5)
/usr/local/go/src/os/exec/exec.go:341 +0x5c
os/exec.(*Cmd).Output(0xc000b60160, 0x11742aa4353, 0x337a580, 0x1, 0x1, 0x106)
/usr/local/go/src/os/exec/exec.go:546 +0x92
github.com/GoogleContainerTools/skaffold/integration/skaffold.(*RunBuilder).RunOrFailOutput(0xc0006bbec0, 0xc0008b9380, 0x0, 0x0, 0x0)
/skaffold/integration/skaffold/helper.go:275 +0x1a9
github.com/GoogleContainerTools/skaffold/integration/skaffold.(*RunBuilder).RunOrFail(...)
/skaffold/integration/skaffold/helper.go:229
github.com/GoogleContainerTools/skaffold/integration.TestDiagnose.func1(0xc0008b9380)
/skaffold/integration/diagnose_test.go:45 +0x1e5
testing.tRunner(0xc0008b9380, 0xc000a5e8d0)
/usr/local/go/src/testing/testing.go:1123 +0xef
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:1168 +0x2b3
goroutine 71 [IO wait, 19 minutes]:
internal/poll.runtime_pollWait(0x7f8e3ce36d18, 0x72, 0x24a18a0)
/usr/local/go/src/runtime/netpoll.go:222 +0x55
internal/poll.(*pollDesc).wait(0xc0007d6718, 0x72, 0x24a1800, 0x32b3ae8, 0x0)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
internal/poll.(*pollDesc).waitRead(...)
/usr/local/go/src/internal/poll/fd_poll_runtime.go:92
internal/poll.(*FD).Read(0xc0007d6700, 0xc000800000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
net.(*netFD).Read(0xc0007d6700, 0xc000800000, 0x1000, 0x1000, 0x43c5dc, 0xc00007cb58, 0x469680)
/usr/local/go/src/net/fd_posix.go:55 +0x4f
net.(*conn).Read(0xc000122e88, 0xc000800000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/net/net.go:182 +0x8e
net/http.(*persistConn).Read(0xc0007f06c0, 0xc000800000, 0x1000, 0x1000, 0xc0007f4240, 0xc00007cc58, 0x407555)
/usr/local/go/src/net/http/transport.go:1887 +0x77
bufio.(*Reader).fill(0xc00075f320)
/usr/local/go/src/bufio/bufio.go:101 +0x105
bufio.(*Reader).Peek(0xc00075f320, 0x1, 0x0, 0x0, 0x1, 0x0, 0xc000745320)
/usr/local/go/src/bufio/bufio.go:139 +0x4f
net/http.(*persistConn).readLoop(0xc0007f06c0)
/usr/local/go/src/net/http/transport.go:2040 +0x1a8
created by net/http.(*Transport).dialConn
/usr/local/go/src/net/http/transport.go:1708 +0xcb7
goroutine 72 [select, 19 minutes]:
net/http.(*persistConn).writeLoop(0xc0007f06c0)
/usr/local/go/src/net/http/transport.go:2340 +0x11c
created by net/http.(*Transport).dialConn
/usr/local/go/src/net/http/transport.go:1709 +0xcdc
FAIL github.com/GoogleContainerTools/skaffold/integration 1200.071s
``` | non_priority | flake testdiagnose bazel times out failure what s odd here is that from the output ran in minutes the test seems to have succeeded but goroutine seems to indicate that it s still executing an external command is it possible this test is running in parallel though wouldn t there be a log statement for the command run testdiagnose bazel time level info msg time level info msg ran in minutes second panic test timed out after goroutine testing m startalarm usr local go src testing testing go created by time gofunc usr local go src time sleep go goroutine testing t run usr local go src testing testing go testing runtests usr local go src testing testing go testing trunner usr local go src testing testing go testing runtests usr local go src testing testing go testing m run usr local go src testing testing go main main testmain go goroutine io klog loggingt flushdaemon root go pkg mod io klog klog go created by io klog init root go pkg mod io klog klog go goroutine github com golang glog loggingt flushdaemon root go pkg mod github com golang glog glog go created by github com golang glog init root go pkg mod github com golang glog glog go goroutine go opencensus io stats view worker start root go pkg mod go opencensus io stats view worker go created by go opencensus io stats view init root go pkg mod go opencensus io stats view worker go goroutine internal poll runtime pollwait usr local go src runtime netpoll go internal poll polldesc wait usr local go src internal poll fd poll runtime go internal poll polldesc waitread usr local go src internal poll fd poll runtime go internal poll fd read usr local go src internal poll fd unix go net netfd read usr local go src net fd posix go net conn read usr local go src net net go crypto tls atleastreader read usr local go src crypto tls conn go bytes buffer readfrom usr local go src bytes buffer go crypto tls conn readfromuntil usr local go src crypto tls conn go crypto tls conn readrecordorccs usr local go src crypto tls conn go crypto tls conn readrecord usr local go src crypto tls conn go crypto tls conn read usr local go src crypto tls conn go bufio reader read usr local go src bufio bufio go io readatleast usr local go src io io go io readfull usr local go src io io go golang org x net readframeheader root go pkg mod golang org x net frame go golang org x net framer readframe root go pkg mod golang org x net frame go golang org x net clientconnreadloop run root go pkg mod golang org x net transport go golang org x net clientconn readloop root go pkg mod golang org x net transport go created by golang org x net transport newclientconn root go pkg mod golang org x net transport go goroutine github com googlecontainertools skaffold pkg skaffold event newhandler skaffold pkg skaffold event event go created by github com googlecontainertools skaffold pkg skaffold event newhandler skaffold pkg skaffold event event go goroutine github com rjeczalik notify nonrecursivetree dispatch root go pkg mod github com rjeczalik notify tree nonrecursive go created by github com rjeczalik notify newnonrecursivetree root go pkg mod github com rjeczalik notify tree nonrecursive go goroutine github com rjeczalik notify nonrecursivetree internal root go pkg mod github com rjeczalik notify tree nonrecursive go created by github com rjeczalik notify newnonrecursivetree root go pkg mod github com rjeczalik notify tree nonrecursive go goroutine testing t run usr local go src testing testing go github com googlecontainertools skaffold integration testdiagnose skaffold integration diagnose test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go goroutine io apimachinery pkg watch streamwatcher receive root go pkg mod io apimachinery pkg watch streamwatcher go created by io apimachinery pkg watch newstreamwatcher root go pkg mod io apimachinery pkg watch streamwatcher go goroutine internal poll runtime pollwait usr local go src runtime netpoll go internal poll polldesc wait usr local go src internal poll fd poll runtime go internal poll polldesc waitread usr local go src internal poll fd poll runtime go internal poll fd read usr local go src internal poll fd unix go os file read usr local go src os file posix go os file read usr local go src os file go bytes buffer readfrom usr local go src bytes buffer go io copybuffer usr local go src io io go io copy usr local go src io io go os exec cmd writerdescriptor usr local go src os exec exec go os exec cmd start usr local go src os exec exec go created by os exec cmd start usr local go src os exec exec go goroutine internal poll runtime pollwait usr local go src runtime netpoll go internal poll polldesc wait usr local go src internal poll fd poll runtime go internal poll polldesc waitread usr local go src internal poll fd poll runtime go internal poll fd read usr local go src internal poll fd unix go os file read usr local go src os file posix go os file read usr local go src os file go io copybuffer usr local go src io io go io copy usr local go src io io go os exec cmd writerdescriptor usr local go src os exec exec go os exec cmd start usr local go src os exec exec go created by os exec cmd start usr local go src os exec exec go goroutine os exec cmd start usr local go src os exec exec go created by os exec cmd start usr local go src os exec exec go goroutine syscall usr local go src syscall asm linux s os process blockuntilwaitable usr local go src os wait waitid go os process wait usr local go src os exec unix go os process wait usr local go src os exec go os exec cmd wait usr local go src os exec exec go os exec cmd run usr local go src os exec exec go os exec cmd output usr local go src os exec exec go github com googlecontainertools skaffold integration skaffold runbuilder runorfailoutput skaffold integration skaffold helper go github com googlecontainertools skaffold integration skaffold runbuilder runorfail skaffold integration skaffold helper go github com googlecontainertools skaffold integration testdiagnose skaffold integration diagnose test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go goroutine internal poll runtime pollwait usr local go src runtime netpoll go internal poll polldesc wait usr local go src internal poll fd poll runtime go internal poll polldesc waitread usr local go src internal poll fd poll runtime go internal poll fd read usr local go src internal poll fd unix go net netfd read usr local go src net fd posix go net conn read usr local go src net net go net http persistconn read usr local go src net http transport go bufio reader fill usr local go src bufio bufio go bufio reader peek usr local go src bufio bufio go net http persistconn readloop usr local go src net http transport go created by net http transport dialconn usr local go src net http transport go goroutine net http persistconn writeloop usr local go src net http transport go created by net http transport dialconn usr local go src net http transport go fail github com googlecontainertools skaffold integration | 0 |
15,633 | 3,968,869,211 | IssuesEvent | 2016-05-03 21:11:51 | california-civic-data-coalition/django-calaccess-raw-data | https://api.github.com/repos/california-civic-data-coalition/django-calaccess-raw-data | closed | Add documentcloud_pages to FilerLinksCd.link_type (in other.py) | documentation enhancement small | ## Overview
Add documentcloud_pages to the **link_type** field on **FilerLinksCd** the database model.
Basically, we think **link_type** should be defined as a "choice field" (in Django parlance), which just means it's a field with a defined list of valid values and an intelligible definition for each value.
Our goal is for every potential choice field to include references to locations in the [official documentation](http://django-calaccess-raw-data.californiacivicdata.org/en/latest/officialdocumentation.html) where the field's valid values and their meanings are described AND to have those choices included in the field's definition.
And we need your help!
## What to do
**Step 1**: Claim this ticket by adding yourself as an **Assignee** (to the right)
**Step 2**: Find where in the [official documentation](http://django-calaccess-raw-data.californiacivicdata.org/en/latest/officialdocumentation.html) link_type's list of valid values is defined.
The best place to look first is in the .CAL layout documents: [.CAL Format version 2.01](http://www.documentcloud.org/documents/2712034-Cal-Format-201.html#document/p1) and [.CAL Format version 1.05.02](http://www.documentcloud.org/documents/2712033-Cal-Format-1-05-02.html#document/p1). These are documents that describe the layout of the .CAL file format, which is the required electronic format of any filings ingested by the CAL-ACCESS system (more on that [here](http://django-calaccess-raw-data.californiacivicdata.org/en/latest/officialdocumentation.html#cal-file-format) if you are interested).
You can start by opening the document and just searching for "link_type". For example, if you were assigned to tackle the stmt_type field on the CvrCampaignDisclosureCd model, you would be looking for something like this::

But note that the same column might appear on multiple tables, and the list of valid values may vary depending on the table. We want to be sure to find the list of valid values for the link_type on the FilerLinksCd model.
Another good place to look is in the [MapCalFormat2Fields](http://www.documentcloud.org/documents/2711616-MapCalFormat2Fields.html#document/p1) document, which describes how .CAL format fields are mapped to CAL-ACCESS database table columns. Note the _Valid_V_ column on the far right.
**Step 3**: Copy the DocumentCloud ID and the page number(s) where link_type's list of valid values is defined.
The DocumentCloud ID can be found the in the URL that points to the document. For example, the url to the .CAL Format version 2.01 doc is `http://www.documentcloud.org/documents/2712034-Cal-Format-201.html#document/p1`, and the id is `2712034-Cal-Format-201`.
And the start page we want to grab is the DocumentCloud page number, not the page number printed on the .PDF. If the list of valid values spans multiple pages, then grab the page where the list ends as well.
**Step 4**: Find where link_type is defined in other.py in the [/calaccess_raw/models/other.py](https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/models/other.py) file underneath the FilerLinksCd class.
Since the model files are rather long, take care to be sure you find the where link_type is defined for the FilerLinksCd model. It should look something like this:
```python
class FilerLinksCd(CalAccessBaseModel):
# ...
# several lines of code later...
# ...
link_type = fields.IntegerField(
max_length=None,
db_column='LINK_TYPE',
blank=False,
verbose_name='link type',
help_text='Denotes the type of the link',
)
```
**Step 5**: Add the document references
Say you found the list of valid values listed in three places. You would list them in the `documentcloud_pages` attribute like this:
```python
class FilerLinksCd(CalAccessBaseModel):
# ...
# several lines of code later...
# ...
link_type = fields.IntegerField(
max_length=None,
db_column='LINK_TYPE',
blank=False,
verbose_name='link type',
help_text='Denotes the type of the link',
documentcloud_pages=[
DocumentCloud(id='2711616-MapCalFormat2Fields', start_page=10),
DocumentCloud(id='2712033-Cal-Format-1-05-02', start_page=19),
DocumentCloud(id='2712034-Cal-Format-201', start_page=24, end_page=25),
]
)
```
Note: You only need to specify the `end_page` if the list spans multiple pages.
**Step 6**: Update the link_type's `help_text` (if necessary)
It might be missing or it might be an empty string or it might not be consistent with what you see in the documentation. Note: The entire help_text string needs to be enclosed in single- or double-quotes like this:
```python
help_text='This is the help text',
```
If the help text is really long (80+ characters), then it needs to span multiple lines like this:
```python
help_text='This is the help text...blah...blah...blah...blah...blah...blah...\
...blah...blah...blah',
```
**If you can,** check the link_type's defined choices against the actual values in link_type's database column.
This will require you to install the app and run the `updatecalaccessrawdata` command. Then you can get the distinct column values with a `GROUP BY` query like this:
```sql
SELECT LINK_TYPE, COUNT(*)
FROM FILER_LINKS_CD
GROUP BY 1
ORDER BY 1;
```
It's likely that you'll find undocumented values in database column, some of which are variants of the valid values. You can note these by repeating the same definition for each variant like this:
```python
OFFICE_CD_CHOICES = (
('GOV', 'Governor'),
('gov', 'Governor'),
('GUV', 'Governor'),
# ...
),
```
If you're dealing with a set of choices that's been consolidated into [calaccess_raw/annotations/choices.py](https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/annotations/choices.py), you can map the different variants to the same definition using the key of the choices dict like this:
```python
OFFICE_CD_CHOICES = (
('gov', look_ups.OFFICE_CODES['GOV']),
('GUV', look_ups.OFFICE_CODES['GOV']),
# ...
),
```
If you're not able to check the database column values for whatever reason, don't sweat it. We'll deal with it in a later issue.
**Wrap up**: Review your changes and create a pull request

That's it!
**PS**: If any of this feels confusing, or if it doesn't seem to match up with what you are seeing, or if you have any suggestions about another approach, don't hesitate to speak up. You can post a comment below, or hit up the #california-civic-data channel on the News Nerdery Slack.
| 1.0 | Add documentcloud_pages to FilerLinksCd.link_type (in other.py) - ## Overview
Add documentcloud_pages to the **link_type** field on **FilerLinksCd** the database model.
Basically, we think **link_type** should be defined as a "choice field" (in Django parlance), which just means it's a field with a defined list of valid values and an intelligible definition for each value.
Our goal is for every potential choice field to include references to locations in the [official documentation](http://django-calaccess-raw-data.californiacivicdata.org/en/latest/officialdocumentation.html) where the field's valid values and their meanings are described AND to have those choices included in the field's definition.
And we need your help!
## What to do
**Step 1**: Claim this ticket by adding yourself as an **Assignee** (to the right)
**Step 2**: Find where in the [official documentation](http://django-calaccess-raw-data.californiacivicdata.org/en/latest/officialdocumentation.html) link_type's list of valid values is defined.
The best place to look first is in the .CAL layout documents: [.CAL Format version 2.01](http://www.documentcloud.org/documents/2712034-Cal-Format-201.html#document/p1) and [.CAL Format version 1.05.02](http://www.documentcloud.org/documents/2712033-Cal-Format-1-05-02.html#document/p1). These are documents that describe the layout of the .CAL file format, which is the required electronic format of any filings ingested by the CAL-ACCESS system (more on that [here](http://django-calaccess-raw-data.californiacivicdata.org/en/latest/officialdocumentation.html#cal-file-format) if you are interested).
You can start by opening the document and just searching for "link_type". For example, if you were assigned to tackle the stmt_type field on the CvrCampaignDisclosureCd model, you would be looking for something like this::

But note that the same column might appear on multiple tables, and the list of valid values may vary depending on the table. We want to be sure to find the list of valid values for the link_type on the FilerLinksCd model.
Another good place to look is in the [MapCalFormat2Fields](http://www.documentcloud.org/documents/2711616-MapCalFormat2Fields.html#document/p1) document, which describes how .CAL format fields are mapped to CAL-ACCESS database table columns. Note the _Valid_V_ column on the far right.
**Step 3**: Copy the DocumentCloud ID and the page number(s) where link_type's list of valid values is defined.
The DocumentCloud ID can be found the in the URL that points to the document. For example, the url to the .CAL Format version 2.01 doc is `http://www.documentcloud.org/documents/2712034-Cal-Format-201.html#document/p1`, and the id is `2712034-Cal-Format-201`.
And the start page we want to grab is the DocumentCloud page number, not the page number printed on the .PDF. If the list of valid values spans multiple pages, then grab the page where the list ends as well.
**Step 4**: Find where link_type is defined in other.py in the [/calaccess_raw/models/other.py](https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/models/other.py) file underneath the FilerLinksCd class.
Since the model files are rather long, take care to be sure you find the where link_type is defined for the FilerLinksCd model. It should look something like this:
```python
class FilerLinksCd(CalAccessBaseModel):
# ...
# several lines of code later...
# ...
link_type = fields.IntegerField(
max_length=None,
db_column='LINK_TYPE',
blank=False,
verbose_name='link type',
help_text='Denotes the type of the link',
)
```
**Step 5**: Add the document references
Say you found the list of valid values listed in three places. You would list them in the `documentcloud_pages` attribute like this:
```python
class FilerLinksCd(CalAccessBaseModel):
# ...
# several lines of code later...
# ...
link_type = fields.IntegerField(
max_length=None,
db_column='LINK_TYPE',
blank=False,
verbose_name='link type',
help_text='Denotes the type of the link',
documentcloud_pages=[
DocumentCloud(id='2711616-MapCalFormat2Fields', start_page=10),
DocumentCloud(id='2712033-Cal-Format-1-05-02', start_page=19),
DocumentCloud(id='2712034-Cal-Format-201', start_page=24, end_page=25),
]
)
```
Note: You only need to specify the `end_page` if the list spans multiple pages.
**Step 6**: Update the link_type's `help_text` (if necessary)
It might be missing or it might be an empty string or it might not be consistent with what you see in the documentation. Note: The entire help_text string needs to be enclosed in single- or double-quotes like this:
```python
help_text='This is the help text',
```
If the help text is really long (80+ characters), then it needs to span multiple lines like this:
```python
help_text='This is the help text...blah...blah...blah...blah...blah...blah...\
...blah...blah...blah',
```
**If you can,** check the link_type's defined choices against the actual values in link_type's database column.
This will require you to install the app and run the `updatecalaccessrawdata` command. Then you can get the distinct column values with a `GROUP BY` query like this:
```sql
SELECT LINK_TYPE, COUNT(*)
FROM FILER_LINKS_CD
GROUP BY 1
ORDER BY 1;
```
It's likely that you'll find undocumented values in database column, some of which are variants of the valid values. You can note these by repeating the same definition for each variant like this:
```python
OFFICE_CD_CHOICES = (
('GOV', 'Governor'),
('gov', 'Governor'),
('GUV', 'Governor'),
# ...
),
```
If you're dealing with a set of choices that's been consolidated into [calaccess_raw/annotations/choices.py](https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/annotations/choices.py), you can map the different variants to the same definition using the key of the choices dict like this:
```python
OFFICE_CD_CHOICES = (
('gov', look_ups.OFFICE_CODES['GOV']),
('GUV', look_ups.OFFICE_CODES['GOV']),
# ...
),
```
If you're not able to check the database column values for whatever reason, don't sweat it. We'll deal with it in a later issue.
**Wrap up**: Review your changes and create a pull request

That's it!
**PS**: If any of this feels confusing, or if it doesn't seem to match up with what you are seeing, or if you have any suggestions about another approach, don't hesitate to speak up. You can post a comment below, or hit up the #california-civic-data channel on the News Nerdery Slack.
| non_priority | add documentcloud pages to filerlinkscd link type in other py overview add documentcloud pages to the link type field on filerlinkscd the database model basically we think link type should be defined as a choice field in django parlance which just means it s a field with a defined list of valid values and an intelligible definition for each value our goal is for every potential choice field to include references to locations in the where the field s valid values and their meanings are described and to have those choices included in the field s definition and we need your help what to do step claim this ticket by adding yourself as an assignee to the right step find where in the link type s list of valid values is defined the best place to look first is in the cal layout documents and these are documents that describe the layout of the cal file format which is the required electronic format of any filings ingested by the cal access system more on that if you are interested you can start by opening the document and just searching for link type for example if you were assigned to tackle the stmt type field on the cvrcampaigndisclosurecd model you would be looking for something like this but note that the same column might appear on multiple tables and the list of valid values may vary depending on the table we want to be sure to find the list of valid values for the link type on the filerlinkscd model another good place to look is in the document which describes how cal format fields are mapped to cal access database table columns note the valid v column on the far right step copy the documentcloud id and the page number s where link type s list of valid values is defined the documentcloud id can be found the in the url that points to the document for example the url to the cal format version doc is and the id is cal format and the start page we want to grab is the documentcloud page number not the page number printed on the pdf if the list of valid values spans multiple pages then grab the page where the list ends as well step find where link type is defined in other py in the file underneath the filerlinkscd class since the model files are rather long take care to be sure you find the where link type is defined for the filerlinkscd model it should look something like this python class filerlinkscd calaccessbasemodel several lines of code later link type fields integerfield max length none db column link type blank false verbose name link type help text denotes the type of the link step add the document references say you found the list of valid values listed in three places you would list them in the documentcloud pages attribute like this python class filerlinkscd calaccessbasemodel several lines of code later link type fields integerfield max length none db column link type blank false verbose name link type help text denotes the type of the link documentcloud pages documentcloud id start page documentcloud id cal format start page documentcloud id cal format start page end page note you only need to specify the end page if the list spans multiple pages step update the link type s help text if necessary it might be missing or it might be an empty string or it might not be consistent with what you see in the documentation note the entire help text string needs to be enclosed in single or double quotes like this python help text this is the help text if the help text is really long characters then it needs to span multiple lines like this python help text this is the help text blah blah blah blah blah blah blah blah blah if you can check the link type s defined choices against the actual values in link type s database column this will require you to install the app and run the updatecalaccessrawdata command then you can get the distinct column values with a group by query like this sql select link type count from filer links cd group by order by it s likely that you ll find undocumented values in database column some of which are variants of the valid values you can note these by repeating the same definition for each variant like this python office cd choices gov governor gov governor guv governor if you re dealing with a set of choices that s been consolidated into you can map the different variants to the same definition using the key of the choices dict like this python office cd choices gov look ups office codes guv look ups office codes if you re not able to check the database column values for whatever reason don t sweat it we ll deal with it in a later issue wrap up review your changes and create a pull request that s it ps if any of this feels confusing or if it doesn t seem to match up with what you are seeing or if you have any suggestions about another approach don t hesitate to speak up you can post a comment below or hit up the california civic data channel on the news nerdery slack | 0 |
603,716 | 18,670,635,277 | IssuesEvent | 2021-10-30 16:38:54 | theseion/Fuel | https://api.github.com/repos/theseion/Fuel | closed | Improve serialization and materialization of Point | auto-migrated Priority-Low Type-Enhancement stale | ```
Idea: If the Point is composed by Integer and they fit a Word, use PointArray.
If they fit in 16 bits, use a ShortPointArray. Otherwise, the regular cluster.
Then for PointArray and ShortPointArray we use the nextWordsPut: that is fast
because of the hackBits.
See PointArrayTest and ShortPointArrayTest,
Other idea is that we can check if the x and y are SmallIntegers and if true,
we diretly use the methods int32: and int32.
```
Original issue reported on code.google.com by `marianopeck` on 12 Sep 2011 at 9:41
| 1.0 | Improve serialization and materialization of Point - ```
Idea: If the Point is composed by Integer and they fit a Word, use PointArray.
If they fit in 16 bits, use a ShortPointArray. Otherwise, the regular cluster.
Then for PointArray and ShortPointArray we use the nextWordsPut: that is fast
because of the hackBits.
See PointArrayTest and ShortPointArrayTest,
Other idea is that we can check if the x and y are SmallIntegers and if true,
we diretly use the methods int32: and int32.
```
Original issue reported on code.google.com by `marianopeck` on 12 Sep 2011 at 9:41
| priority | improve serialization and materialization of point idea if the point is composed by integer and they fit a word use pointarray if they fit in bits use a shortpointarray otherwise the regular cluster then for pointarray and shortpointarray we use the nextwordsput that is fast because of the hackbits see pointarraytest and shortpointarraytest other idea is that we can check if the x and y are smallintegers and if true we diretly use the methods and original issue reported on code google com by marianopeck on sep at | 1 |
16,707 | 3,550,222,768 | IssuesEvent | 2016-01-20 21:08:20 | tgstation/-tg-station | https://api.github.com/repos/tgstation/-tg-station | closed | You can interact with things while asleep. | Bug In Game Exploit Needs Reproducing/Testing Priority: High | I was attacked by a ice spider, the sleep toxin put me to sleep and soon after another spider came by to feast on my innards. Luckily, I was able to somehow beat him to death even though I was unconscious and I was also able to shake my fellow crew member up from his slumber as well. | 1.0 | You can interact with things while asleep. - I was attacked by a ice spider, the sleep toxin put me to sleep and soon after another spider came by to feast on my innards. Luckily, I was able to somehow beat him to death even though I was unconscious and I was also able to shake my fellow crew member up from his slumber as well. | non_priority | you can interact with things while asleep i was attacked by a ice spider the sleep toxin put me to sleep and soon after another spider came by to feast on my innards luckily i was able to somehow beat him to death even though i was unconscious and i was also able to shake my fellow crew member up from his slumber as well | 0 |
177,839 | 6,587,754,456 | IssuesEvent | 2017-09-13 22:30:42 | ZenHubIO/support | https://api.github.com/repos/ZenHubIO/support | closed | ToDo List needs a "Back" Button | Enhancement Low Priority ToDo Feature | #### Description
##### User Story
As a user, when I navigate to my ToDo list in ZH I should have the option to navigate back to my last GH page without having to use my browser's back button. If I navigate from board --> ToDo then the ToDo page should offer a "back to [XYZ] board" button. If I navigate from milestones --> ToDo then the ToDo button should offer a "back to [repo name] milestones" button. ...and so on.
| 1.0 | ToDo List needs a "Back" Button - #### Description
##### User Story
As a user, when I navigate to my ToDo list in ZH I should have the option to navigate back to my last GH page without having to use my browser's back button. If I navigate from board --> ToDo then the ToDo page should offer a "back to [XYZ] board" button. If I navigate from milestones --> ToDo then the ToDo button should offer a "back to [repo name] milestones" button. ...and so on.
| priority | todo list needs a back button description user story as a user when i navigate to my todo list in zh i should have the option to navigate back to my last gh page without having to use my browser s back button if i navigate from board todo then the todo page should offer a back to board button if i navigate from milestones todo then the todo button should offer a back to milestones button and so on | 1 |
49,548 | 6,224,169,341 | IssuesEvent | 2017-07-10 13:44:01 | fabric8-ui/fabric8-ux | https://api.github.com/repos/fabric8-ui/fabric8-ux | opened | [VISUALS] Experimental Feature Animation - Concept | concept visual design | Verification Conditions:
- Try creating a subtle animation for the experimental features pages
- Share out the animation for feedback
- Collect and incorporate feedback
- If we decide to proceed with the animation, deliver assets to dev | 1.0 | [VISUALS] Experimental Feature Animation - Concept - Verification Conditions:
- Try creating a subtle animation for the experimental features pages
- Share out the animation for feedback
- Collect and incorporate feedback
- If we decide to proceed with the animation, deliver assets to dev | non_priority | experimental feature animation concept verification conditions try creating a subtle animation for the experimental features pages share out the animation for feedback collect and incorporate feedback if we decide to proceed with the animation deliver assets to dev | 0 |
283,633 | 30,913,506,446 | IssuesEvent | 2023-08-05 02:05:52 | hshivhare67/kernel_v4.19.72 | https://api.github.com/repos/hshivhare67/kernel_v4.19.72 | reopened | CVE-2023-2235 (High) detected in linuxlinux-4.19.282 | Mend: dependency security vulnerability | ## CVE-2023-2235 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use-after-free vulnerability in the Linux Kernel Performance Events system can be exploited to achieve local privilege escalation.
The perf_group_detach function did not check the event's siblings' attach_state before calling add_event_to_groups(), but remove_on_exec made it possible to call list_del_event() on before detaching from their group, making it possible to use a dangling pointer causing a use-after-free vulnerability.
We recommend upgrading past commit fd0815f632c24878e325821943edccc7fde947a2.
<p>Publish Date: 2023-05-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-2235>CVE-2023-2235</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-2235">https://www.linuxkernelcves.com/cves/CVE-2023-2235</a></p>
<p>Release Date: 2023-05-01</p>
<p>Fix Resolution: v5.15.104,v6.1.21,v6.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-2235 (High) detected in linuxlinux-4.19.282 - ## CVE-2023-2235 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use-after-free vulnerability in the Linux Kernel Performance Events system can be exploited to achieve local privilege escalation.
The perf_group_detach function did not check the event's siblings' attach_state before calling add_event_to_groups(), but remove_on_exec made it possible to call list_del_event() on before detaching from their group, making it possible to use a dangling pointer causing a use-after-free vulnerability.
We recommend upgrading past commit fd0815f632c24878e325821943edccc7fde947a2.
<p>Publish Date: 2023-05-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-2235>CVE-2023-2235</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-2235">https://www.linuxkernelcves.com/cves/CVE-2023-2235</a></p>
<p>Release Date: 2023-05-01</p>
<p>Fix Resolution: v5.15.104,v6.1.21,v6.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details a use after free vulnerability in the linux kernel performance events system can be exploited to achieve local privilege escalation the perf group detach function did not check the event s siblings attach state before calling add event to groups but remove on exec made it possible to call list del event on before detaching from their group making it possible to use a dangling pointer causing a use after free vulnerability we recommend upgrading past commit publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
379,062 | 11,214,911,718 | IssuesEvent | 2020-01-07 00:06:40 | octobercms/october | https://api.github.com/repos/octobercms/october | closed | Missing settings for InlineStyle button in Froala editor | Priority: Low Status: Archived Type: Conceptual Enhancement help wanted | If I use "inlineStyle" button in Froala, there is "Big red" and "Small blue" css class to select in editor toolbar.
I think, setting for "inlineStyle" button is missing in "Settings > Editor settings > Markup classes" | 1.0 | Missing settings for InlineStyle button in Froala editor - If I use "inlineStyle" button in Froala, there is "Big red" and "Small blue" css class to select in editor toolbar.
I think, setting for "inlineStyle" button is missing in "Settings > Editor settings > Markup classes" | priority | missing settings for inlinestyle button in froala editor if i use inlinestyle button in froala there is big red and small blue css class to select in editor toolbar i think setting for inlinestyle button is missing in settings editor settings markup classes | 1 |
567,968 | 16,943,284,761 | IssuesEvent | 2021-06-28 00:14:13 | DoobDev/Doob | https://api.github.com/repos/DoobDev/Doob | closed | Make it so you don't have to move files to run Doob. (#193) | High Priority Stale development | >this is most likely possible by catching the error being thrown when you can login to stuff like topgg and statcord. | 1.0 | Make it so you don't have to move files to run Doob. (#193) - >this is most likely possible by catching the error being thrown when you can login to stuff like topgg and statcord. | priority | make it so you don t have to move files to run doob this is most likely possible by catching the error being thrown when you can login to stuff like topgg and statcord | 1 |
346,787 | 10,419,837,618 | IssuesEvent | 2019-09-15 19:30:38 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Aghosting breaks the character menu | Priority: 1-high Type: Bug | Repro steps:
1. Enter `aghost` into the debug console or `/aghost` into the chat window
2. Repeat step one to exit aghost mode.
3. Open the character menu with "c". It should be broken, only displaying a title bar with an empty window. | 1.0 | Aghosting breaks the character menu - Repro steps:
1. Enter `aghost` into the debug console or `/aghost` into the chat window
2. Repeat step one to exit aghost mode.
3. Open the character menu with "c". It should be broken, only displaying a title bar with an empty window. | priority | aghosting breaks the character menu repro steps enter aghost into the debug console or aghost into the chat window repeat step one to exit aghost mode open the character menu with c it should be broken only displaying a title bar with an empty window | 1 |
513,389 | 14,920,858,911 | IssuesEvent | 2021-01-23 07:07:12 | takapy0210/paper | https://api.github.com/repos/takapy0210/paper | closed | 150 Successful Machine Learning Models: 6 Lessons Learned at Booking.com | 2021/01 Booking.com paper priority:high | ## Meta
- Link: https://dl.acm.org/doi/pdf/10.1145/3292500.3330744
- Author: Booking.com
- Date:
- Journal: KDD19
## どんなもの?(3行ぐらいで)
- 機械学習を取り入れる中で失敗と成功の両方を何度も経験してきたBooking.comが、150の成功モデルから学んだことについての論文
- Booking.com特有の下記問題に対して機械学習を用いてアプローチしてきた
- High Stakes
- Infinitesimal Queries
- Complex Items
- Constrained Supply
- Continuous Cold Start
- Content Overload
- そして成功から学んだ6つの教訓を紹介している
## 先行研究と比べてどこがすごい?
- 機械学習が商業的な利用においてどのような影響を及ぼすかを論じたものはほとんどなく、そこについて論じている点
## 技術や手法のキモはどこ?
1. 機械学習モデルを導入したプロジェクトは大きなビジネス価値をもたらす
2. 機械学習モデルのパフォーマンスは、ビジネスのパフォーマンスと同じにならない
3. 解決しようとする問題をはっきりさせることが大切
4. 予測は遅延をもたらす
5. モデルの質に関して迅速なフィードバックを得ること
6. ランダム化比較試験を用いてモデルを使った時のビジネスインパクトを測る
### 1. 機械学習モデルを導入したプロジェクトは大きなビジネス価値をもたらす
Booking.comが使う機械学習モデルは、大きく以下の6つのカテゴリ
- Traveller Preference Models: 旅行者の好みを幅広く予測するモデル
- 行き先や日程などについて、どれくらい柔軟に考えているか(こだわっているか)推定するモデル
- Traveller Context Models: 旅行の文脈を予測するモデル(a)
- カップルなのか、家族なのか、車で近くに行きたいのか飛行機で遠くに行きたいかなどのコンテクストを推定するモデル
- Item Space Navigation Models: 履歴などからユーザーの見た内容をトラッキングするモデル
- スクロール、クリック、ソート、フィルタなどの暗黙フィードバックからユーザーの好みを推定するモデル
- User Interface Optimization Models: UIを最適化するモデル
- リスト内の項目数、背景色や画像などをコンテキストとユーザー情報を考慮して最適化するモデル
- Content Curation: レビューを始めとするコンテンツを集めてどれを表示するか決めるモデル(b)
- レビューをキュレーションして、簡易で代表的な要約を構築するモデル
- Content Augmentation: 価格やオプションのトレンドをコンピューティングするモデル
- Great Value: 同程度のオプションの宿と比べて価格が良いものを通知するモデル
- Price Trends: 時期や目的地によって価格は変動するので、ユーザーが旅行をしたい日の価格を予測し、傾向があればそれをユーザーに通知するためのモデル。これは宿泊施設ではなく、目的地が通知される(c)

上記のカテゴリで、ビジネス的な価値を生んでいる。
Content Curationだけはベンチマークを下回る結果になっている。
(ベンチマークの詳細は分からなかったので、ちょっとモヤモヤ)

### 2. 機械学習モデルのパフォーマンスは、ビジネスのパフォーマンスと同じにならない
Booking.comではランダム化比較試験を用いることで、機械学習モデルによって提供されたものの価値を分析しているが、モデルのパフォーマンスが上がることは、必ずしもビジネス価値を上げるとはいえないことが分かった。
これは、ビジネスの価値が既に飽和していたり、そもそも提供されるものの利用者が少なかったり、Uncanny Valley effect現象(下図)が生じたりと、さまざまな理由がある。

(Booking.comのユーザーが「私はザルツブルクとロンドンについて入力しただけなのに、なぜBooking.comは私がウィーンに行こうとしていることまで知っているのだ?」とユーザーが不思議に思っている様子)
### 3. 解決しようとする問題をはっきりさせることが大切
機械学習モデルを作る際には、問題を明確にすることが非常に重要になります。パフォーマンスの向上は、モデルの改良ではなく、解くタスクそのものを変えることによってもたらされることもある。
### 4. 予測は遅延をもたらす
Booking.comで人工的に遅延を起こす実験をしたところ、レイテンシが30%増加すると、コンバージョン率が0.5%下がることが示された。よりパラメーターが少ないモデルや自社製の線形予測エンジンなどを使うことでレイテンシの長さを最小化することが重要。(遅延と予測精度のバランスが大事ね)
### 5. モデルの質に関して迅速なフィードバックを得ること
機械学習モデルがリクエストに正しく反応しているのかは、出力のクオリティをモニタリングして検証する必要がある。しかし、多くの場合は真のラベルを観察できなかったり、予測を行ってから数日後、あるいは数週間後に真のラベルが観測できるパターンもあるという部分が課題としてあった。
この点について、Booking.comは二項分類でモデルが出力した応答の分布を観察し、品質をチェックしている。

### 6. RCTを用いてモデルを使った時のビジネスインパクトを測る
成功した機械学習モデルの多くは、高度な実験デザインのもとで行われていた、とBooking.comは述べています。
論文には、複数のタイプのランダム化比較試験に関する提案が含まれています。ランダム化比較試験において統制群は変更が施されていないプラットフォームにアクセスし、実験群だけが変更が施されたプラットフォームにアクセスします。例えば、全てのユーザーに変更を露出するのが適切ではない場合は、適格なサブセットから統制群と、実験群を作ります。
| 1.0 | 150 Successful Machine Learning Models: 6 Lessons Learned at Booking.com - ## Meta
- Link: https://dl.acm.org/doi/pdf/10.1145/3292500.3330744
- Author: Booking.com
- Date:
- Journal: KDD19
## どんなもの?(3行ぐらいで)
- 機械学習を取り入れる中で失敗と成功の両方を何度も経験してきたBooking.comが、150の成功モデルから学んだことについての論文
- Booking.com特有の下記問題に対して機械学習を用いてアプローチしてきた
- High Stakes
- Infinitesimal Queries
- Complex Items
- Constrained Supply
- Continuous Cold Start
- Content Overload
- そして成功から学んだ6つの教訓を紹介している
## 先行研究と比べてどこがすごい?
- 機械学習が商業的な利用においてどのような影響を及ぼすかを論じたものはほとんどなく、そこについて論じている点
## 技術や手法のキモはどこ?
1. 機械学習モデルを導入したプロジェクトは大きなビジネス価値をもたらす
2. 機械学習モデルのパフォーマンスは、ビジネスのパフォーマンスと同じにならない
3. 解決しようとする問題をはっきりさせることが大切
4. 予測は遅延をもたらす
5. モデルの質に関して迅速なフィードバックを得ること
6. ランダム化比較試験を用いてモデルを使った時のビジネスインパクトを測る
### 1. 機械学習モデルを導入したプロジェクトは大きなビジネス価値をもたらす
Booking.comが使う機械学習モデルは、大きく以下の6つのカテゴリ
- Traveller Preference Models: 旅行者の好みを幅広く予測するモデル
- 行き先や日程などについて、どれくらい柔軟に考えているか(こだわっているか)推定するモデル
- Traveller Context Models: 旅行の文脈を予測するモデル(a)
- カップルなのか、家族なのか、車で近くに行きたいのか飛行機で遠くに行きたいかなどのコンテクストを推定するモデル
- Item Space Navigation Models: 履歴などからユーザーの見た内容をトラッキングするモデル
- スクロール、クリック、ソート、フィルタなどの暗黙フィードバックからユーザーの好みを推定するモデル
- User Interface Optimization Models: UIを最適化するモデル
- リスト内の項目数、背景色や画像などをコンテキストとユーザー情報を考慮して最適化するモデル
- Content Curation: レビューを始めとするコンテンツを集めてどれを表示するか決めるモデル(b)
- レビューをキュレーションして、簡易で代表的な要約を構築するモデル
- Content Augmentation: 価格やオプションのトレンドをコンピューティングするモデル
- Great Value: 同程度のオプションの宿と比べて価格が良いものを通知するモデル
- Price Trends: 時期や目的地によって価格は変動するので、ユーザーが旅行をしたい日の価格を予測し、傾向があればそれをユーザーに通知するためのモデル。これは宿泊施設ではなく、目的地が通知される(c)

上記のカテゴリで、ビジネス的な価値を生んでいる。
Content Curationだけはベンチマークを下回る結果になっている。
(ベンチマークの詳細は分からなかったので、ちょっとモヤモヤ)

### 2. 機械学習モデルのパフォーマンスは、ビジネスのパフォーマンスと同じにならない
Booking.comではランダム化比較試験を用いることで、機械学習モデルによって提供されたものの価値を分析しているが、モデルのパフォーマンスが上がることは、必ずしもビジネス価値を上げるとはいえないことが分かった。
これは、ビジネスの価値が既に飽和していたり、そもそも提供されるものの利用者が少なかったり、Uncanny Valley effect現象(下図)が生じたりと、さまざまな理由がある。

(Booking.comのユーザーが「私はザルツブルクとロンドンについて入力しただけなのに、なぜBooking.comは私がウィーンに行こうとしていることまで知っているのだ?」とユーザーが不思議に思っている様子)
### 3. 解決しようとする問題をはっきりさせることが大切
機械学習モデルを作る際には、問題を明確にすることが非常に重要になります。パフォーマンスの向上は、モデルの改良ではなく、解くタスクそのものを変えることによってもたらされることもある。
### 4. 予測は遅延をもたらす
Booking.comで人工的に遅延を起こす実験をしたところ、レイテンシが30%増加すると、コンバージョン率が0.5%下がることが示された。よりパラメーターが少ないモデルや自社製の線形予測エンジンなどを使うことでレイテンシの長さを最小化することが重要。(遅延と予測精度のバランスが大事ね)
### 5. モデルの質に関して迅速なフィードバックを得ること
機械学習モデルがリクエストに正しく反応しているのかは、出力のクオリティをモニタリングして検証する必要がある。しかし、多くの場合は真のラベルを観察できなかったり、予測を行ってから数日後、あるいは数週間後に真のラベルが観測できるパターンもあるという部分が課題としてあった。
この点について、Booking.comは二項分類でモデルが出力した応答の分布を観察し、品質をチェックしている。

### 6. RCTを用いてモデルを使った時のビジネスインパクトを測る
成功した機械学習モデルの多くは、高度な実験デザインのもとで行われていた、とBooking.comは述べています。
論文には、複数のタイプのランダム化比較試験に関する提案が含まれています。ランダム化比較試験において統制群は変更が施されていないプラットフォームにアクセスし、実験群だけが変更が施されたプラットフォームにアクセスします。例えば、全てのユーザーに変更を露出するのが適切ではない場合は、適格なサブセットから統制群と、実験群を作ります。
| priority | successful machine learning models lessons learned at booking com meta link author booking com date journal どんなもの?( ) 機械学習を取り入れる中で失敗と成功の両方を何度も経験してきたbooking comが、 booking com特有の下記問題に対して機械学習を用いてアプローチしてきた high stakes infinitesimal queries complex items constrained supply continuous cold start content overload 先行研究と比べてどこがすごい? 機械学習が商業的な利用においてどのような影響を及ぼすかを論じたものはほとんどなく、そこについて論じている点 技術や手法のキモはどこ? 機械学習モデルを導入したプロジェクトは大きなビジネス価値をもたらす 機械学習モデルのパフォーマンスは、ビジネスのパフォーマンスと同じにならない 解決しようとする問題をはっきりさせることが大切 予測は遅延をもたらす モデルの質に関して迅速なフィードバックを得ること ランダム化比較試験を用いてモデルを使った時のビジネスインパクトを測る 機械学習モデルを導入したプロジェクトは大きなビジネス価値をもたらす booking comが使う機械学習モデルは、 traveller preference models 旅行者の好みを幅広く予測するモデル 行き先や日程などについて、どれくらい柔軟に考えているか(こだわっているか)推定するモデル traveller context models 旅行の文脈を予測するモデル(a) カップルなのか、家族なのか、車で近くに行きたいのか飛行機で遠くに行きたいかなどのコンテクストを推定するモデル item space navigation models 履歴などからユーザーの見た内容をトラッキングするモデル スクロール、クリック、ソート、フィルタなどの暗黙フィードバックからユーザーの好みを推定するモデル user interface optimization models uiを最適化するモデル リスト内の項目数、背景色や画像などをコンテキストとユーザー情報を考慮して最適化するモデル content curation レビューを始めとするコンテンツを集めてどれを表示するか決めるモデル(b) レビューをキュレーションして、簡易で代表的な要約を構築するモデル content augmentation 価格やオプションのトレンドをコンピューティングするモデル great value 同程度のオプションの宿と比べて価格が良いものを通知するモデル price trends 時期や目的地によって価格は変動するので、ユーザーが旅行をしたい日の価格を予測し、傾向があればそれをユーザーに通知するためのモデル。これは宿泊施設ではなく、目的地が通知される(c) 上記のカテゴリで、ビジネス的な価値を生んでいる。 content curationだけはベンチマークを下回る結果になっている。 (ベンチマークの詳細は分からなかったので、ちょっとモヤモヤ) 機械学習モデルのパフォーマンスは、ビジネスのパフォーマンスと同じにならない booking comではランダム化比較試験を用いることで、機械学習モデルによって提供されたものの価値を分析しているが、モデルのパフォーマンスが上がることは、必ずしもビジネス価値を上げるとはいえないことが分かった。 これは、ビジネスの価値が既に飽和していたり、そもそも提供されるものの利用者が少なかったり、uncanny valley effect現象(下図)が生じたりと、さまざまな理由がある。 (booking comのユーザーが「私はザルツブルクとロンドンについて入力しただけなのに、なぜbooking comは私がウィーンに行こうとしていることまで知っているのだ?」とユーザーが不思議に思っている様子) 解決しようとする問題をはっきりさせることが大切 機械学習モデルを作る際には、問題を明確にすることが非常に重要になります。パフォーマンスの向上は、モデルの改良ではなく、解くタスクそのものを変えることによってもたらされることもある。 予測は遅延をもたらす booking comで人工的に遅延を起こす実験をしたところ、 %増加すると、 %下がることが示された。よりパラメーターが少ないモデルや自社製の線形予測エンジンなどを使うことでレイテンシの長さを最小化することが重要。(遅延と予測精度のバランスが大事ね) モデルの質に関して迅速なフィードバックを得ること 機械学習モデルがリクエストに正しく反応しているのかは、出力のクオリティをモニタリングして検証する必要がある。しかし、多くの場合は真のラベルを観察できなかったり、予測を行ってから数日後、あるいは数週間後に真のラベルが観測できるパターンもあるという部分が課題としてあった。 この点について、booking comは二項分類でモデルが出力した応答の分布を観察し、品質をチェックしている。 rctを用いてモデルを使った時のビジネスインパクトを測る 成功した機械学習モデルの多くは、高度な実験デザインのもとで行われていた、とbooking comは述べています。 論文には、複数のタイプのランダム化比較試験に関する提案が含まれています。ランダム化比較試験において統制群は変更が施されていないプラットフォームにアクセスし、実験群だけが変更が施されたプラットフォームにアクセスします。例えば、全てのユーザーに変更を露出するのが適切ではない場合は、適格なサブセットから統制群と、実験群を作ります。 | 1 |
783,435 | 27,530,565,078 | IssuesEvent | 2023-03-06 21:43:22 | spwalgren/Student-Budget-Tracker | https://api.github.com/repos/spwalgren/Student-Budget-Tracker | closed | Linkup Frontend with Backend for Transaction System | enhancement High Priority | The transaction system on the frontend currently uses data stored in the frontend. It should make API calls to the backend and use data from the database. | 1.0 | Linkup Frontend with Backend for Transaction System - The transaction system on the frontend currently uses data stored in the frontend. It should make API calls to the backend and use data from the database. | priority | linkup frontend with backend for transaction system the transaction system on the frontend currently uses data stored in the frontend it should make api calls to the backend and use data from the database | 1 |
278,689 | 8,649,078,281 | IssuesEvent | 2018-11-26 18:19:34 | kubernetes-sigs/cluster-api-provider-aws | https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-aws | closed | make manifests-dev sometimes fails | kind/bug priority/important-longterm | /kind bug
I was really surprised to find inconsistent behavior here:
```
salazar:cluster-api-provider-aws cha$ make manifests-dev
MANAGER_IMAGE=gcr.io/chuck-heptio/cluster-api-aws-controller:0.0.2 /Applications/Xcode.app/Contents/Developer/usr/bin/make manifests
./cmd/clusterctl/examples/aws/generate-yaml.sh
Done generating /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/cluster.yaml
Done generating /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/machines.yaml
Done generating /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/aws_manager_image_patch.yaml
Done copying /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/addons.yaml
clusterawsadm alpha bootstrap generate-aws-default-profile > cmd/clusterctl/examples/aws/out/credentials
no such flag -v
make[1]: *** [cmd/clusterctl/examples/aws/out/credentials] Error 1
make: *** [manifests-dev] Error 2
salazar:cluster-api-provider-aws cha$ make manifests-dev
MANAGER_IMAGE=gcr.io/chuck-heptio/cluster-api-aws-controller:0.0.2 /Applications/Xcode.app/Contents/Developer/usr/bin/make manifests
go run vendor/sigs.k8s.io/controller-tools/cmd/controller-gen/main.go crd
CRD files generated, files can be found under path /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/config/crds.
kustomize build config/default/ > cmd/clusterctl/examples/aws/out/provider-components.yaml
2018/11/16 14:01:44 Adding nameprefix to Namespace resource will be deprecated in next release.
echo "---" >> cmd/clusterctl/examples/aws/out/provider-components.yaml
kustomize build vendor/sigs.k8s.io/cluster-api/config/default/ >> cmd/clusterctl/examples/aws/out/provider-components.yaml
2018/11/16 14:01:44 Adding nameprefix to Namespace resource will be deprecated in next release.
salazar:cluster-api-provider-aws cha$ make manifests-dev
MANAGER_IMAGE=gcr.io/chuck-heptio/cluster-api-aws-controller:0.0.2 /Applications/Xcode.app/Contents/Developer/usr/bin/make manifests
clusterawsadm alpha bootstrap generate-aws-default-profile > cmd/clusterctl/examples/aws/out/credentials
no such flag -v
make[1]: *** [cmd/clusterctl/examples/aws/out/credentials] Error 1
make: *** [manifests-dev] Error 2
salazar:cluster-api-provider-aws cha$
```
This is three runs, it fails twice and succeeds once.
**What steps did you take and what happened:**
I ran `make manifests-dev` a few times in a row.
**What did you expect to happen:**
I expected the same thing each time.
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
**Environment:**
- Cluster-api-provider-aws version:
- Kubernetes version: (use `kubectl version`):
- OS (e.g. from `/etc/os-release`): | 1.0 | make manifests-dev sometimes fails - /kind bug
I was really surprised to find inconsistent behavior here:
```
salazar:cluster-api-provider-aws cha$ make manifests-dev
MANAGER_IMAGE=gcr.io/chuck-heptio/cluster-api-aws-controller:0.0.2 /Applications/Xcode.app/Contents/Developer/usr/bin/make manifests
./cmd/clusterctl/examples/aws/generate-yaml.sh
Done generating /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/cluster.yaml
Done generating /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/machines.yaml
Done generating /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/aws_manager_image_patch.yaml
Done copying /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/cmd/clusterctl/examples/aws/out/addons.yaml
clusterawsadm alpha bootstrap generate-aws-default-profile > cmd/clusterctl/examples/aws/out/credentials
no such flag -v
make[1]: *** [cmd/clusterctl/examples/aws/out/credentials] Error 1
make: *** [manifests-dev] Error 2
salazar:cluster-api-provider-aws cha$ make manifests-dev
MANAGER_IMAGE=gcr.io/chuck-heptio/cluster-api-aws-controller:0.0.2 /Applications/Xcode.app/Contents/Developer/usr/bin/make manifests
go run vendor/sigs.k8s.io/controller-tools/cmd/controller-gen/main.go crd
CRD files generated, files can be found under path /Users/cha/go/src/sigs.k8s.io/cluster-api-provider-aws/config/crds.
kustomize build config/default/ > cmd/clusterctl/examples/aws/out/provider-components.yaml
2018/11/16 14:01:44 Adding nameprefix to Namespace resource will be deprecated in next release.
echo "---" >> cmd/clusterctl/examples/aws/out/provider-components.yaml
kustomize build vendor/sigs.k8s.io/cluster-api/config/default/ >> cmd/clusterctl/examples/aws/out/provider-components.yaml
2018/11/16 14:01:44 Adding nameprefix to Namespace resource will be deprecated in next release.
salazar:cluster-api-provider-aws cha$ make manifests-dev
MANAGER_IMAGE=gcr.io/chuck-heptio/cluster-api-aws-controller:0.0.2 /Applications/Xcode.app/Contents/Developer/usr/bin/make manifests
clusterawsadm alpha bootstrap generate-aws-default-profile > cmd/clusterctl/examples/aws/out/credentials
no such flag -v
make[1]: *** [cmd/clusterctl/examples/aws/out/credentials] Error 1
make: *** [manifests-dev] Error 2
salazar:cluster-api-provider-aws cha$
```
This is three runs, it fails twice and succeeds once.
**What steps did you take and what happened:**
I ran `make manifests-dev` a few times in a row.
**What did you expect to happen:**
I expected the same thing each time.
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
**Environment:**
- Cluster-api-provider-aws version:
- Kubernetes version: (use `kubectl version`):
- OS (e.g. from `/etc/os-release`): | priority | make manifests dev sometimes fails kind bug i was really surprised to find inconsistent behavior here salazar cluster api provider aws cha make manifests dev manager image gcr io chuck heptio cluster api aws controller applications xcode app contents developer usr bin make manifests cmd clusterctl examples aws generate yaml sh done generating users cha go src sigs io cluster api provider aws cmd clusterctl examples aws out cluster yaml done generating users cha go src sigs io cluster api provider aws cmd clusterctl examples aws out machines yaml done generating users cha go src sigs io cluster api provider aws cmd clusterctl examples aws out aws manager image patch yaml done copying users cha go src sigs io cluster api provider aws cmd clusterctl examples aws out addons yaml clusterawsadm alpha bootstrap generate aws default profile cmd clusterctl examples aws out credentials no such flag v make error make error salazar cluster api provider aws cha make manifests dev manager image gcr io chuck heptio cluster api aws controller applications xcode app contents developer usr bin make manifests go run vendor sigs io controller tools cmd controller gen main go crd crd files generated files can be found under path users cha go src sigs io cluster api provider aws config crds kustomize build config default cmd clusterctl examples aws out provider components yaml adding nameprefix to namespace resource will be deprecated in next release echo cmd clusterctl examples aws out provider components yaml kustomize build vendor sigs io cluster api config default cmd clusterctl examples aws out provider components yaml adding nameprefix to namespace resource will be deprecated in next release salazar cluster api provider aws cha make manifests dev manager image gcr io chuck heptio cluster api aws controller applications xcode app contents developer usr bin make manifests clusterawsadm alpha bootstrap generate aws default profile cmd clusterctl examples aws out credentials no such flag v make error make error salazar cluster api provider aws cha this is three runs it fails twice and succeeds once what steps did you take and what happened i ran make manifests dev a few times in a row what did you expect to happen i expected the same thing each time anything else you would like to add environment cluster api provider aws version kubernetes version use kubectl version os e g from etc os release | 1 |
451,154 | 13,025,597,305 | IssuesEvent | 2020-07-27 13:46:57 | HEXRD/hexrdgui | https://api.github.com/repos/HEXRD/hexrdgui | closed | Materials table not updating on new instrument configuration load | bug high priority | Just noticed this one. Loaded two configurations with different X-ray wavelengths, and noticed that the materials table (specifically the two-theta values) don't get recalculated. I can force the update by reloading the materials database, but we need to get the wavelength-based update of the `planeData` object wired in. The class will auto update all of the energy-dependent parameters by setting the `planeData.wavelength` object with the energy in keV (I know, pathological...). | 1.0 | Materials table not updating on new instrument configuration load - Just noticed this one. Loaded two configurations with different X-ray wavelengths, and noticed that the materials table (specifically the two-theta values) don't get recalculated. I can force the update by reloading the materials database, but we need to get the wavelength-based update of the `planeData` object wired in. The class will auto update all of the energy-dependent parameters by setting the `planeData.wavelength` object with the energy in keV (I know, pathological...). | priority | materials table not updating on new instrument configuration load just noticed this one loaded two configurations with different x ray wavelengths and noticed that the materials table specifically the two theta values don t get recalculated i can force the update by reloading the materials database but we need to get the wavelength based update of the planedata object wired in the class will auto update all of the energy dependent parameters by setting the planedata wavelength object with the energy in kev i know pathological | 1 |
319,394 | 9,743,271,575 | IssuesEvent | 2019-06-03 00:34:37 | inf3rno/e2e | https://api.github.com/repos/inf3rno/e2e | opened | Check automated and manual browsing differences | do: experiment on: library priority 3: usual | We need to do some experimentation to know what can be automated and what not. For example I am curious if HTML5 validation for form inputs still works if we set the values programmatically. An even more interesting thing is if we should set the values and fire events or just set the values and events will be fired by the browser automagically. We need to document the results too because the library will be developed based on that info. | 1.0 | Check automated and manual browsing differences - We need to do some experimentation to know what can be automated and what not. For example I am curious if HTML5 validation for form inputs still works if we set the values programmatically. An even more interesting thing is if we should set the values and fire events or just set the values and events will be fired by the browser automagically. We need to document the results too because the library will be developed based on that info. | priority | check automated and manual browsing differences we need to do some experimentation to know what can be automated and what not for example i am curious if validation for form inputs still works if we set the values programmatically an even more interesting thing is if we should set the values and fire events or just set the values and events will be fired by the browser automagically we need to document the results too because the library will be developed based on that info | 1 |
301,368 | 26,041,914,276 | IssuesEvent | 2022-12-22 11:09:16 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | reopened | [Flaky Test] Should render dynamic blocks when the meta box uses the excerpt for front end rendering | [Feature] Meta Boxes [Status] Stale [Type] Flaky Test | <!-- __META_DATA__:{"failedTimes":1,"totalCommits":1,"baseCommit":"6bae9d7b153dd052cdeb88d23bf87ded31bae098"} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
Should render dynamic blocks when the meta box uses the excerpt for front end rendering
## Test path
`specs/editor/plugins/meta-boxes.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><time datetime="2021-11-03T01:14:18.982Z"><code>[2021-11-03T01:14:18.982Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1414772163"><code>update/mark-post-navigation-link-and-term-description-as-stable</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2021-11-08T13:32:31.991Z"><code>[2021-11-08T13:32:31.991Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1434869385"><code>polish/navigation-setup-state</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2021-11-23T13:33:56.008Z"><code>[2021-11-23T13:33:56.008Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1494646686"><code>try/navigation-remove-all-pages</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-01-03T17:56:10.718Z"><code>[2022-01-03T17:56:10.718Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1650319465"><code>rnmobile/update/refactor-editor-initialization</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-01-12T08:16:47.576Z"><code>[2022-01-12T08:16:47.576Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1686487885"><code>fix/improve-group-control-backfrop-render-animation</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-01-24T10:05:42.431Z"><code>[2022-01-24T10:05:42.431Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1739105171"><code>try/check-link-color-in-contrast-checker</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-04-07T08:22:38.374Z"><code>[2022-04-07T08:22:38.374Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2107484344"><code>trunk</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-05-04T03:22:43.049Z"><code>[2022-05-04T03:22:43.049Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2267489986"><code>try/use-css-var-for-user-presets</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-09-01T08:41:57.111Z"><code>[2022-09-01T08:41:57.111Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2969986736"><code>migreate/query-test</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2022-11-07T14:38:48.411Z"><code>[2022-11-07T14:38:48.411Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/3411083158"><code>try/site-editor-subtle-outlines</code></a>.
</summary>
```
● Meta boxes › Should render dynamic blocks when the meta box uses the excerpt for front end rendering
TimeoutError: waiting for function failed: timeout 30000ms exceeded
at new WaitTask (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:813:28)
at DOMWorld.waitForFunction (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:728:22)
at Frame.waitForFunction (../../node_modules/puppeteer-core/src/common/FrameManager.ts:1351:28)
at Page.waitForFunction (../../node_modules/puppeteer-core/src/common/Page.ts:3346:29)
at waitForInserterCloseAndContentFocus (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:94:8)
at insertBlock (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:173:8)
at runMicrotasks (<anonymous>)
at Object.<anonymous> (specs/editor/plugins/meta-boxes.test.js:52:3)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| 1.0 | [Flaky Test] Should render dynamic blocks when the meta box uses the excerpt for front end rendering - <!-- __META_DATA__:{"failedTimes":1,"totalCommits":1,"baseCommit":"6bae9d7b153dd052cdeb88d23bf87ded31bae098"} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
Should render dynamic blocks when the meta box uses the excerpt for front end rendering
## Test path
`specs/editor/plugins/meta-boxes.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><time datetime="2021-11-03T01:14:18.982Z"><code>[2021-11-03T01:14:18.982Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1414772163"><code>update/mark-post-navigation-link-and-term-description-as-stable</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2021-11-08T13:32:31.991Z"><code>[2021-11-08T13:32:31.991Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1434869385"><code>polish/navigation-setup-state</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2021-11-23T13:33:56.008Z"><code>[2021-11-23T13:33:56.008Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1494646686"><code>try/navigation-remove-all-pages</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-01-03T17:56:10.718Z"><code>[2022-01-03T17:56:10.718Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1650319465"><code>rnmobile/update/refactor-editor-initialization</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-01-12T08:16:47.576Z"><code>[2022-01-12T08:16:47.576Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1686487885"><code>fix/improve-group-control-backfrop-render-animation</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-01-24T10:05:42.431Z"><code>[2022-01-24T10:05:42.431Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/1739105171"><code>try/check-link-color-in-contrast-checker</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-04-07T08:22:38.374Z"><code>[2022-04-07T08:22:38.374Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2107484344"><code>trunk</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-05-04T03:22:43.049Z"><code>[2022-05-04T03:22:43.049Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2267489986"><code>try/use-css-var-for-user-presets</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><time datetime="2022-09-01T08:41:57.111Z"><code>[2022-09-01T08:41:57.111Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2969986736"><code>migreate/query-test</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2022-11-07T14:38:48.411Z"><code>[2022-11-07T14:38:48.411Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/3411083158"><code>try/site-editor-subtle-outlines</code></a>.
</summary>
```
● Meta boxes › Should render dynamic blocks when the meta box uses the excerpt for front end rendering
TimeoutError: waiting for function failed: timeout 30000ms exceeded
at new WaitTask (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:813:28)
at DOMWorld.waitForFunction (../../node_modules/puppeteer-core/src/common/DOMWorld.ts:728:22)
at Frame.waitForFunction (../../node_modules/puppeteer-core/src/common/FrameManager.ts:1351:28)
at Page.waitForFunction (../../node_modules/puppeteer-core/src/common/Page.ts:3346:29)
at waitForInserterCloseAndContentFocus (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:94:8)
at insertBlock (../e2e-test-utils/build/@wordpress/e2e-test-utils/src/inserter.js:173:8)
at runMicrotasks (<anonymous>)
at Object.<anonymous> (specs/editor/plugins/meta-boxes.test.js:52:3)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| non_priority | should render dynamic blocks when the meta box uses the excerpt for front end rendering flaky test detected this is an auto generated issue by github actions please do not edit this manually test title should render dynamic blocks when the meta box uses the excerpt for front end rendering test path specs editor plugins meta boxes test js errors test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on test passed after failed attempt on a href ● meta boxes › should render dynamic blocks when the meta box uses the excerpt for front end rendering timeouterror waiting for function failed timeout exceeded at new waittask node modules puppeteer core src common domworld ts at domworld waitforfunction node modules puppeteer core src common domworld ts at frame waitforfunction node modules puppeteer core src common framemanager ts at page waitforfunction node modules puppeteer core src common page ts at waitforinsertercloseandcontentfocus test utils build wordpress test utils src inserter js at insertblock test utils build wordpress test utils src inserter js at runmicrotasks at object specs editor plugins meta boxes test js | 0 |
600,701 | 18,351,944,747 | IssuesEvent | 2021-10-08 13:33:08 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | opened | Gantt duplicates requests when reordering items in the TreeList part | Bug SEV: High C: Gantt jQuery2 Priority 5 | ### Bug report
When tasks are reordered in Gantt duplicated requests are sent for updating.
### Reproduction of the problem
1. Open the [Gantt Overview Demo ](https://demos.telerik.com/kendo-ui/gantt/index)
2. Reorder two tasks and inspect the requests in the Network tab
### Current behavior
For example, in case two tasks on the same level are changing their places, four requests are sent, two by two the requests are the same.
In case a task is moved on a different level multiple requests are sent, as all are duplicated.
### Expected/desired behavior
There should be no duplicated requests when reordering tasks in Gantt
This is a regression introduced with R3 2020 (2020.3.915)
### Environment
* **Kendo UI version:** 2021.3.914
* **Browser:** [all ]
| 1.0 | Gantt duplicates requests when reordering items in the TreeList part - ### Bug report
When tasks are reordered in Gantt duplicated requests are sent for updating.
### Reproduction of the problem
1. Open the [Gantt Overview Demo ](https://demos.telerik.com/kendo-ui/gantt/index)
2. Reorder two tasks and inspect the requests in the Network tab
### Current behavior
For example, in case two tasks on the same level are changing their places, four requests are sent, two by two the requests are the same.
In case a task is moved on a different level multiple requests are sent, as all are duplicated.
### Expected/desired behavior
There should be no duplicated requests when reordering tasks in Gantt
This is a regression introduced with R3 2020 (2020.3.915)
### Environment
* **Kendo UI version:** 2021.3.914
* **Browser:** [all ]
| priority | gantt duplicates requests when reordering items in the treelist part bug report when tasks are reordered in gantt duplicated requests are sent for updating reproduction of the problem open the reorder two tasks and inspect the requests in the network tab current behavior for example in case two tasks on the same level are changing their places four requests are sent two by two the requests are the same in case a task is moved on a different level multiple requests are sent as all are duplicated expected desired behavior there should be no duplicated requests when reordering tasks in gantt this is a regression introduced with environment kendo ui version browser | 1 |
6,103 | 3,331,724,929 | IssuesEvent | 2015-11-11 16:59:06 | phetsims/gravity-and-orbits | https://api.github.com/repos/phetsims/gravity-and-orbits | reopened | anti-pattern for abstract types | code review | Related to code review #173.
ModeConfig is intended to be an abstract base type, so the first thing would be to describe it as such in the JSdoc at the top of the file.
Then there's this function that must be implemented by subtypes:
```js
// abstract
/**
* @returns {Array<BodyConfiguration}
*/
getBodies: function() {}
```
This pattern is not recommended because (in the general case) it can succeed silently when it should fail.
The more typical pattern for functions in abstract types is to either (a) not define them so that they fail, or (b) implement a function that fails. My preference is for (b), since it also provides the opportunity to document. So recommended to replace the above with:
```
/**
* @returns {BodyConfiguration[]}
* @public
* @abstract
*/
getBodies: function() {
throw new Error( 'must be implemented by subtype' );
}
``` | 1.0 | anti-pattern for abstract types - Related to code review #173.
ModeConfig is intended to be an abstract base type, so the first thing would be to describe it as such in the JSdoc at the top of the file.
Then there's this function that must be implemented by subtypes:
```js
// abstract
/**
* @returns {Array<BodyConfiguration}
*/
getBodies: function() {}
```
This pattern is not recommended because (in the general case) it can succeed silently when it should fail.
The more typical pattern for functions in abstract types is to either (a) not define them so that they fail, or (b) implement a function that fails. My preference is for (b), since it also provides the opportunity to document. So recommended to replace the above with:
```
/**
* @returns {BodyConfiguration[]}
* @public
* @abstract
*/
getBodies: function() {
throw new Error( 'must be implemented by subtype' );
}
``` | non_priority | anti pattern for abstract types related to code review modeconfig is intended to be an abstract base type so the first thing would be to describe it as such in the jsdoc at the top of the file then there s this function that must be implemented by subtypes js abstract returns array bodyconfiguration getbodies function this pattern is not recommended because in the general case it can succeed silently when it should fail the more typical pattern for functions in abstract types is to either a not define them so that they fail or b implement a function that fails my preference is for b since it also provides the opportunity to document so recommended to replace the above with returns bodyconfiguration public abstract getbodies function throw new error must be implemented by subtype | 0 |
396,476 | 11,709,643,467 | IssuesEvent | 2020-03-08 20:01:11 | bozcani/borsa | https://api.github.com/repos/bozcani/borsa | closed | Create a page to show stock data update status | enhancement priority: high | It is difficult to follow which stocks are up-to-date. Let's create a web page for tracking the status.
Requirements:
- List stocks in a table
- Add columns: ticker symbol, market, the last update.
- Add a button to update OHLCV data of stock. | 1.0 | Create a page to show stock data update status - It is difficult to follow which stocks are up-to-date. Let's create a web page for tracking the status.
Requirements:
- List stocks in a table
- Add columns: ticker symbol, market, the last update.
- Add a button to update OHLCV data of stock. | priority | create a page to show stock data update status it is difficult to follow which stocks are up to date let s create a web page for tracking the status requirements list stocks in a table add columns ticker symbol market the last update add a button to update ohlcv data of stock | 1 |
255,968 | 19,348,586,404 | IssuesEvent | 2021-12-15 13:31:16 | open-telemetry/opentelemetry-collector-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib | closed | Contribution guideline - add new component to tests | documentation good first issue | New components are expected to be enabled in the related component-specific tests like done in this PR:
https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/6517
We do not seem to have this documented anywhere at this moment and we should. This improvement to the docs should come accompanied by reference PRs for new components. For instance, a new extension could have its first PR similar to this:
https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/6634 | 1.0 | Contribution guideline - add new component to tests - New components are expected to be enabled in the related component-specific tests like done in this PR:
https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/6517
We do not seem to have this documented anywhere at this moment and we should. This improvement to the docs should come accompanied by reference PRs for new components. For instance, a new extension could have its first PR similar to this:
https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/6634 | non_priority | contribution guideline add new component to tests new components are expected to be enabled in the related component specific tests like done in this pr we do not seem to have this documented anywhere at this moment and we should this improvement to the docs should come accompanied by reference prs for new components for instance a new extension could have its first pr similar to this | 0 |
724,091 | 24,917,151,465 | IssuesEvent | 2022-10-30 14:39:07 | kubernetes/ingress-nginx | https://api.github.com/repos/kubernetes/ingress-nginx | closed | Changing .Values.controller.name results in a certificate not valid for domain error | kind/support lifecycle/rotten needs-triage needs-priority |
<!--
Welcome to ingress-nginx! For a smooth issue process, try to answer the following questions.
Don't worry if they're not all applicable; just try to include what you can :-)
If you need to include code snippets or logs, please put them in fenced code
blocks. If they're super-long, please use the details tag like
<details><summary>super-long log</summary> lots of stuff </details>
-->
<!--
IMPORTANT!!!
Please complete the next sections or the issue will be closed.
This questions are the first thing we need to know to understand the context.
-->
**NGINX Ingress controller version** (exec into the pod and run nginx-ingress-controller --version.):
<!--
POD_NAMESPACE=ingress-nginx
POD_NAME=$(kubectl get pods -n $POD_NAMESPACE -l app.kubernetes.io/name=ingress-nginx --field-selector=status.phase=Running -o jsonpath='{.items[0].metadata.name}')
kubectl exec -it $POD_NAME -n $POD_NAMESPACE -- /nginx-ingress-controller --version
-->
**Kubernetes version** (use `kubectl version`):
**Environment**:
- **Cloud provider or hardware configuration**:
Oracle virtual box
- **OS** (e.g. from /etc/os-release):
```
NAME="Ubuntu"
VERSION="20.04.4 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.4 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal
```
- **Kernel** (e.g. `uname -a`):
```
Linux alienware 5.13.0-44-generic #49~20.04.1-Ubuntu SMP Wed May 18 18:44:28 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
```
- **Install tools**:
```
Seen this issue in both a kubeadm and minukube created clusters
```
- **Basic cluster related info**:
- `kubectl version`
```
WARNING: This version information is deprecated and will be replaced with the output from kubectl version --short. Use --output=yaml|json to get the full version.
Client Version: version.Info{Major:"1", Minor:"24", GitVersion:"v1.24.0", GitCommit:"4ce5a8954017644c5420bae81d72b09b735c21f0", GitTreeState:"clean", BuildDate:"2022-05-04T02:28:17Z", GoVersion:"go1.18.1", Compiler:"gc", Platform:"linux/amd64"}
Kustomize Version: v4.5.4
Server Version: version.Info{Major:"1", Minor:"21", GitVersion:"v1.21.1", GitCommit:"5e58841cce77d4bc13713ad2b91fa0d961e69192", GitTreeState:"clean", BuildDate:"2021-05-21T23:01:33Z", GoVersion:"go1.16.4", Compiler:"gc", Platform:"linux/amd64"}
WARNING: version difference between client (1.24) and server (1.21)
```
- `kubectl get nodes -o wide`
```
kubectl get nodes -o wide
NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME
kind-control-plane Ready control-plane,master 94d v1.21.1 172.18.0.2 <none> Ubuntu 21.04 5.13.0-44-generic containerd://1.5.2
```
- **How was the ingress-nginx-controller installed**:
- If helm was used then please show output of `helm ls -A | grep -i ingress`
```
helm ls -A | grep -i ingress
ic-ingress-nginx kube-system 1 2022-05-30 18:12:27.287105885 -0300 -03 deployed ingress-nginx-4.0.15 1.1.1
```
- If helm was used then please show output of `helm -n <ingresscontrollernamepspace> get values <helmreleasename>`
```
helm -n kube-system get values ic-ingress-nginx
USER-SUPPLIED VALUES:
controller:
name: notcontroller
service:
ipFamilies:
- IPv4
```
- **Current State of the controller**:
- `kubectl describe ingressclasses`
```
kubectl describe ingressclasses
Name: nginx
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=ingress-nginx
app.kubernetes.io/version=1.1.1
helm.sh/chart=ingress-nginx-4.0.15
Annotations: meta.helm.sh/release-name: ic-ingress-nginx
meta.helm.sh/release-namespace: kube-system
Controller: k8s.io/ingress-nginx
Events: <none>
```
- `kubectl -n <ingresscontrollernamespace> get all -A -o wide`
```
kubectl -n kube-system get all -A -o wide
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
default pod/apple-app 1/1 Running 0 3m5s 10.244.0.12 kind-control-plane <none> <none>
default pod/banana-app 1/1 Running 0 3m5s 10.244.0.13 kind-control-plane <none> <none>
kube-system pod/coredns-558bd4d5db-pw9xj 1/1 Running 6 95d 10.244.0.5 kind-control-plane <none> <none>
kube-system pod/coredns-558bd4d5db-xxsfx 1/1 Running 6 95d 10.244.0.6 kind-control-plane <none> <none>
kube-system pod/etcd-kind-control-plane 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/ic-ingress-nginx-notcontroller-67f9445d98-8tgm8 1/1 Running 0 3m14s 10.244.0.10 kind-control-plane <none> <none>
kube-system pod/kindnet-2qw5t 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-apiserver-kind-control-plane 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-controller-manager-kind-control-plane 1/1 Running 7 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-proxy-z7r75 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-scheduler-kind-control-plane 1/1 Running 7 95d 172.18.0.2 kind-control-plane <none> <none>
local-path-storage pod/local-path-provisioner-547f784dff-7rcxd 1/1 Running 11 95d 10.244.0.4 kind-control-plane <none> <none>
NAMESPACE NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE SELECTOR
default service/apple-service ClusterIP 10.96.118.150 <none> 5678/TCP 3m5s app=apple
default service/banana-service ClusterIP 10.96.155.84 <none> 5678/TCP 3m5s app=banana
default service/kubernetes ClusterIP 10.96.0.1 <none> 443/TCP 95d <none>
kube-system service/ic-ingress-nginx-notcontroller LoadBalancer 10.96.187.204 <pending> 80:30496/TCP,443:31975/TCP 3m14s app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
kube-system service/ic-ingress-nginx-notcontroller-admission ClusterIP 10.96.24.137 <none> 443/TCP 3m14s app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
kube-system service/kube-dns ClusterIP 10.96.0.10 <none> 53/UDP,53/TCP,9153/TCP 95d k8s-app=kube-dns
NAMESPACE NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE CONTAINERS IMAGES SELECTOR
kube-system daemonset.apps/kindnet 1 1 1 1 1 <none> 95d kindnet-cni docker.io/kindest/kindnetd:v20210326-1e038dc5 app=kindnet
kube-system daemonset.apps/kube-proxy 1 1 1 1 1 kubernetes.io/os=linux 95d kube-proxy k8s.gcr.io/kube-proxy:v1.21.1 k8s-app=kube-proxy
NAMESPACE NAME READY UP-TO-DATE AVAILABLE AGE CONTAINERS IMAGES SELECTOR
kube-system deployment.apps/coredns 2/2 2 2 95d coredns k8s.gcr.io/coredns/coredns:v1.8.0 k8s-app=kube-dns
kube-system deployment.apps/ic-ingress-nginx-notcontroller 1/1 1 1 3m14s controller k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
local-path-storage deployment.apps/local-path-provisioner 1/1 1 1 95d local-path-provisioner docker.io/rancher/local-path-provisioner:v0.0.14 app=local-path-provisioner
NAMESPACE NAME DESIRED CURRENT READY AGE CONTAINERS IMAGES SELECTOR
kube-system replicaset.apps/coredns-558bd4d5db 2 2 2 95d coredns k8s.gcr.io/coredns/coredns:v1.8.0 k8s-app=kube-dns,pod-template-hash=558bd4d5db
kube-system replicaset.apps/ic-ingress-nginx-notcontroller-67f9445d98 1 1 1 3m14s controller k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx,pod-template-hash=67f9445d98
local-path-storage replicaset.apps/local-path-provisioner-547f784dff 1 1 1 95d local-path-provisioner docker.io/rancher/local-path-provisioner:v0.0.14 app=local-path-provisioner,pod-template-hash=547f784dff
```
- `kubectl -n <ingresscontrollernamespace> describe po <ingresscontrollerpodname>`
```
kubectl -n kube-system describe po ic-ingress-nginx-notcontroller-67f9445d98-8tgm8
Name: ic-ingress-nginx-notcontroller-67f9445d98-8tgm8
Namespace: kube-system
Priority: 0
Node: kind-control-plane/172.18.0.2
Start Time: Wed, 01 Jun 2022 16:35:40 -0300
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/name=ingress-nginx
pod-template-hash=67f9445d98
Annotations: <none>
Status: Running
IP: 10.244.0.10
IPs:
IP: 10.244.0.10
Controlled By: ReplicaSet/ic-ingress-nginx-notcontroller-67f9445d98
Containers:
controller:
Container ID: containerd://5a4c0f5ed21200b6b82e0afebd175076056100a5235f92f8c0287d0055956c7e
Image: k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de
Image ID: k8s.gcr.io/ingress-nginx/controller@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de
Ports: 80/TCP, 443/TCP, 8443/TCP
Host Ports: 0/TCP, 0/TCP, 0/TCP
Args:
/nginx-ingress-controller
--publish-service=$(POD_NAMESPACE)/ic-ingress-nginx-notcontroller
--election-id=ingress-controller-leader
--controller-class=k8s.io/ingress-nginx
--configmap=$(POD_NAMESPACE)/ic-ingress-nginx-notcontroller
--validating-webhook=:8443
--validating-webhook-certificate=/usr/local/certificates/cert
--validating-webhook-key=/usr/local/certificates/key
State: Running
Started: Wed, 01 Jun 2022 16:35:41 -0300
Ready: True
Restart Count: 0
Requests:
cpu: 100m
memory: 90Mi
Liveness: http-get http://:10254/healthz delay=10s timeout=1s period=10s #success=1 #failure=5
Readiness: http-get http://:10254/healthz delay=10s timeout=1s period=10s #success=1 #failure=3
Environment:
POD_NAME: ic-ingress-nginx-notcontroller-67f9445d98-8tgm8 (v1:metadata.name)
POD_NAMESPACE: kube-system (v1:metadata.namespace)
LD_PRELOAD: /usr/local/lib/libmimalloc.so
Mounts:
/usr/local/certificates/ from webhook-cert (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4pcbr (ro)
Conditions:
Type Status
Initialized True
Ready True
ContainersReady True
PodScheduled True
Volumes:
webhook-cert:
Type: Secret (a volume populated by a Secret)
SecretName: ic-ingress-nginx-admission
Optional: false
kube-api-access-4pcbr:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
QoS Class: Burstable
Node-Selectors: kubernetes.io/os=linux
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 4m18s default-scheduler Successfully assigned kube-system/ic-ingress-nginx-notcontroller-67f9445d98-8tgm8 to kind-control-plane
Normal Pulled 4m18s kubelet Container image "k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de" already present on machine
Normal Created 4m18s kubelet Created container controller
Normal Started 4m18s kubelet Started container controller
Normal RELOAD 4m16s nginx-ingress-controller NGINX reload triggered due to a change in configuration
```
- `kubectl -n <ingresscontrollernamespace> describe svc <ingresscontrollerservicename>`
```
kubectl -n kube-system describe svc ic-ingress-nginx-notcontroller
Name: ic-ingress-nginx-notcontroller
Namespace: kube-system
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=ingress-nginx
app.kubernetes.io/version=1.1.1
helm.sh/chart=ingress-nginx-4.0.15
Annotations: meta.helm.sh/release-name: ic-ingress-nginx
meta.helm.sh/release-namespace: kube-system
Selector: app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
Type: LoadBalancer
IP Family Policy: SingleStack
IP Families: IPv4
IP: 10.96.187.204
IPs: 10.96.187.204
Port: http 80/TCP
TargetPort: http/TCP
NodePort: http 30496/TCP
Endpoints: 10.244.0.10:80
Port: https 443/TCP
TargetPort: https/TCP
NodePort: https 31975/TCP
Endpoints: 10.244.0.10:443
Session Affinity: None
External Traffic Policy: Cluster
Events: <none>
kubectl -n kube-system describe svc ic-ingress-nginx-notcontroller-admission
Name: ic-ingress-nginx-notcontroller-admission
Namespace: kube-system
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=ingress-nginx
app.kubernetes.io/version=1.1.1
helm.sh/chart=ingress-nginx-4.0.15
Annotations: meta.helm.sh/release-name: ic-ingress-nginx
meta.helm.sh/release-namespace: kube-system
Selector: app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
Type: ClusterIP
IP Family Policy: SingleStack
IP Families: IPv4
IP: 10.96.24.137
IPs: 10.96.24.137
Port: https-webhook 443/TCP
TargetPort: webhook/TCP
Endpoints: 10.244.0.10:8443
Session Affinity: None
Events: <none>
```
- **Current state of ingress object, if applicable**:
- `kubectl -n <appnnamespace> get all,ing -o wide`
```
kubectl get all,ing -o wide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
pod/apple-app 1/1 Running 0 7m7s 10.244.0.12 kind-control-plane <none> <none>
pod/banana-app 1/1 Running 0 7m7s 10.244.0.13 kind-control-plane <none> <none>
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE SELECTOR
service/apple-service ClusterIP 10.96.118.150 <none> 5678/TCP 7m7s app=apple
service/banana-service ClusterIP 10.96.155.84 <none> 5678/TCP 7m7s app=banana
service/kubernetes ClusterIP 10.96.0.1 <none> 443/TCP 95d <none>
```
- Snippet used to apply this app and ingress resource:
```
cat <<EOF> simple-ingress.yml
kind: Pod
apiVersion: v1
metadata:
name: apple-app
labels:
app: apple
spec:
containers:
- name: apple-app
image: hashicorp/http-echo
args:
- "-text=apple"
---
kind: Service
apiVersion: v1
metadata:
name: apple-service
spec:
selector:
app: apple
ports:
- port: 5678 # Default port for image
---
kind: Pod
apiVersion: v1
metadata:
name: banana-app
labels:
app: banana
spec:
containers:
- name: banana-app
image: hashicorp/http-echo
args:
- "-text=banana"
---
kind: Service
apiVersion: v1
metadata:
name: banana-service
spec:
selector:
app: banana
ports:
- port: 5678 # Default port for image
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: example-ingress
annotations:
ingress.kubernetes.io/rewrite-target: /
spec:
ingressClassName: nginx
rules:
- http:
paths:
- backend:
service:
name: apple-service
port:
number: 5678
path: /apple
pathType: ImplementationSpecific
- backend:
service:
name: banana-service
port:
number: 5678
path: /banana
pathType: ImplementationSpecific
EOF
kubectl apply -f simple-ingress.yml
```
**What happened**:
It seems like .Values.controller.name cannot be different than 'controller'. Changing it to someting else results in an certificate not valid for domain error:
```
kubectl apply -f simple-ingress.yml
pod/apple-app unchanged
service/apple-service unchanged
pod/banana-app unchanged
service/banana-service unchanged
Error from server (InternalError): error when creating "simple-ingress.yml": Internal error occurred: failed calling webhook "validate.nginx.ingress.kubernetes.io": Post "https://ic-ingress-nginx-notcontroller-admission.kube-system.svc:443/networking/v1/ingresses?timeout=10s": x509: certificate is valid for ic-ingress-nginx-controller-admission, ic-ingress-nginx-controller-admission.kube-system.svc, not ic-ingress-nginx-notcontroller-admission.kube-system.svc
```
**What you expected to happen**:
The Ingress resource should be created without domain errors.
**How to reproduce it**:
* create the values file
```
cat <<EOF> values.yml
controller:
service:
ipFamilies:
- IPv4
EOF
```
* install version 4.0.15 using the previously created values.yml file
```
helm install ic-ingress-nginx ingress-nginx/ingress-nginx --version 4.0.15 -n kube-system -f values.yml
```
* create the ingress resource ( sucessfully this time )
```
cat <<EOF> simple-ingress.yml
kind: Pod
apiVersion: v1
metadata:
name: apple-app
labels:
app: apple
spec:
containers:
- name: apple-app
image: hashicorp/http-echo
args:
- "-text=apple"
---
kind: Service
apiVersion: v1
metadata:
name: apple-service
spec:
selector:
app: apple
ports:
- port: 5678 # Default port for image
---
kind: Pod
apiVersion: v1
metadata:
name: banana-app
labels:
app: banana
spec:
containers:
- name: banana-app
image: hashicorp/http-echo
args:
- "-text=banana"
---
kind: Service
apiVersion: v1
metadata:
name: banana-service
spec:
selector:
app: banana
ports:
- port: 5678 # Default port for image
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: example-ingress
annotations:
ingress.kubernetes.io/rewrite-target: /
spec:
ingressClassName: nginx
rules:
- http:
paths:
- backend:
service:
name: apple-service
port:
number: 5678
path: /apple
pathType: ImplementationSpecific
- backend:
service:
name: banana-service
port:
number: 5678
path: /banana
pathType: ImplementationSpecific
EOF
kubectl apply -f simple-ingress.yml
```
* Check that it works
```
POD_NAME=$(kubectl get pods -n kube-system -l app.kubernetes.io/name=ingress-nginx -o NAME)
kubectl exec -it -n kube-system $POD_NAME -- curl -H 'Host: foo.bar' localhost/banana
```
* Edit values.yml file and add an override for .Values.controller.name to something other than controller
```
cat <<EOF> values.yml
controller:
name: notcontroller
service:
ipFamilies:
- IPv4
EOF
```
* Run helm update using that upgrade values.yml file
```
helm upgrade ic-ingress-nginx ingress-nginx/ingress-nginx --version 4.0.15 -n kube-system -f values.yml
```
* Delete the resources of simple-ingress.yml
```
kubectl delete -f simple-ingress.yml
```
* Create the resources defined in simple-ingress.yml file again ( This time the ERROR happens )
```
kubectl apply -f simple-ingress.yml
pod/apple-app created
service/apple-service created
pod/banana-app created
service/banana-service created
Error from server (InternalError): error when creating "simple-ingress.yml": Internal error occurred: failed calling webhook "validate.nginx.ingress.kubernetes.io": Post "https://ic-ingress-nginx-notcontroller-admission.kube-system.svc:443/networking/v1/ingresses?timeout=10s": x509: certificate is valid for ic-ingress-nginx-controller-admission, ic-ingress-nginx-controller-admission.kube-system.svc, not ic-ingress-nginx-notcontroller-admission.kube-system.svc
```
<!---
As minimally and precisely as possible. Keep in mind we do not have access to your cluster or application.
Help up us (if possible) reproducing the issue using minikube or kind.
## Install minikube/kind
- Minikube https://minikube.sigs.k8s.io/docs/start/
- Kind https://kind.sigs.k8s.io/docs/user/quick-start/
## Install the ingress controller
kubectl apply -f https://raw.githubusercontent.com/kubernetes/ingress-nginx/main/deploy/static/provider/baremetal/deploy.yaml
## Install an application that will act as default backend (is just an echo app)
kubectl apply -f https://raw.githubusercontent.com/kubernetes/ingress-nginx/main/docs/examples/http-svc.yaml
## Create an ingress (please add any additional annotation required)
echo "
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: foo-bar
annotations:
kubernetes.io/ingress.class: nginx
spec:
ingressClassName: nginx # omit this if you're on controller version below 1.0.0
rules:
- host: foo.bar
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: http-svc
port:
number: 80
" | kubectl apply -f -
## make a request
POD_NAME=$(k get pods -n ingress-nginx -l app.kubernetes.io/name=ingress-nginx -o NAME)
kubectl exec -it -n ingress-nginx $POD_NAME -- curl -H 'Host: foo.bar' localhost
--->
<!-- If this is actually about documentation, uncomment the following block -->
<!--
/kind documentation
/remove-kind bug
-->
| 1.0 | Changing .Values.controller.name results in a certificate not valid for domain error -
<!--
Welcome to ingress-nginx! For a smooth issue process, try to answer the following questions.
Don't worry if they're not all applicable; just try to include what you can :-)
If you need to include code snippets or logs, please put them in fenced code
blocks. If they're super-long, please use the details tag like
<details><summary>super-long log</summary> lots of stuff </details>
-->
<!--
IMPORTANT!!!
Please complete the next sections or the issue will be closed.
This questions are the first thing we need to know to understand the context.
-->
**NGINX Ingress controller version** (exec into the pod and run nginx-ingress-controller --version.):
<!--
POD_NAMESPACE=ingress-nginx
POD_NAME=$(kubectl get pods -n $POD_NAMESPACE -l app.kubernetes.io/name=ingress-nginx --field-selector=status.phase=Running -o jsonpath='{.items[0].metadata.name}')
kubectl exec -it $POD_NAME -n $POD_NAMESPACE -- /nginx-ingress-controller --version
-->
**Kubernetes version** (use `kubectl version`):
**Environment**:
- **Cloud provider or hardware configuration**:
Oracle virtual box
- **OS** (e.g. from /etc/os-release):
```
NAME="Ubuntu"
VERSION="20.04.4 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.4 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal
```
- **Kernel** (e.g. `uname -a`):
```
Linux alienware 5.13.0-44-generic #49~20.04.1-Ubuntu SMP Wed May 18 18:44:28 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
```
- **Install tools**:
```
Seen this issue in both a kubeadm and minukube created clusters
```
- **Basic cluster related info**:
- `kubectl version`
```
WARNING: This version information is deprecated and will be replaced with the output from kubectl version --short. Use --output=yaml|json to get the full version.
Client Version: version.Info{Major:"1", Minor:"24", GitVersion:"v1.24.0", GitCommit:"4ce5a8954017644c5420bae81d72b09b735c21f0", GitTreeState:"clean", BuildDate:"2022-05-04T02:28:17Z", GoVersion:"go1.18.1", Compiler:"gc", Platform:"linux/amd64"}
Kustomize Version: v4.5.4
Server Version: version.Info{Major:"1", Minor:"21", GitVersion:"v1.21.1", GitCommit:"5e58841cce77d4bc13713ad2b91fa0d961e69192", GitTreeState:"clean", BuildDate:"2021-05-21T23:01:33Z", GoVersion:"go1.16.4", Compiler:"gc", Platform:"linux/amd64"}
WARNING: version difference between client (1.24) and server (1.21)
```
- `kubectl get nodes -o wide`
```
kubectl get nodes -o wide
NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME
kind-control-plane Ready control-plane,master 94d v1.21.1 172.18.0.2 <none> Ubuntu 21.04 5.13.0-44-generic containerd://1.5.2
```
- **How was the ingress-nginx-controller installed**:
- If helm was used then please show output of `helm ls -A | grep -i ingress`
```
helm ls -A | grep -i ingress
ic-ingress-nginx kube-system 1 2022-05-30 18:12:27.287105885 -0300 -03 deployed ingress-nginx-4.0.15 1.1.1
```
- If helm was used then please show output of `helm -n <ingresscontrollernamepspace> get values <helmreleasename>`
```
helm -n kube-system get values ic-ingress-nginx
USER-SUPPLIED VALUES:
controller:
name: notcontroller
service:
ipFamilies:
- IPv4
```
- **Current State of the controller**:
- `kubectl describe ingressclasses`
```
kubectl describe ingressclasses
Name: nginx
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=ingress-nginx
app.kubernetes.io/version=1.1.1
helm.sh/chart=ingress-nginx-4.0.15
Annotations: meta.helm.sh/release-name: ic-ingress-nginx
meta.helm.sh/release-namespace: kube-system
Controller: k8s.io/ingress-nginx
Events: <none>
```
- `kubectl -n <ingresscontrollernamespace> get all -A -o wide`
```
kubectl -n kube-system get all -A -o wide
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
default pod/apple-app 1/1 Running 0 3m5s 10.244.0.12 kind-control-plane <none> <none>
default pod/banana-app 1/1 Running 0 3m5s 10.244.0.13 kind-control-plane <none> <none>
kube-system pod/coredns-558bd4d5db-pw9xj 1/1 Running 6 95d 10.244.0.5 kind-control-plane <none> <none>
kube-system pod/coredns-558bd4d5db-xxsfx 1/1 Running 6 95d 10.244.0.6 kind-control-plane <none> <none>
kube-system pod/etcd-kind-control-plane 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/ic-ingress-nginx-notcontroller-67f9445d98-8tgm8 1/1 Running 0 3m14s 10.244.0.10 kind-control-plane <none> <none>
kube-system pod/kindnet-2qw5t 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-apiserver-kind-control-plane 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-controller-manager-kind-control-plane 1/1 Running 7 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-proxy-z7r75 1/1 Running 6 95d 172.18.0.2 kind-control-plane <none> <none>
kube-system pod/kube-scheduler-kind-control-plane 1/1 Running 7 95d 172.18.0.2 kind-control-plane <none> <none>
local-path-storage pod/local-path-provisioner-547f784dff-7rcxd 1/1 Running 11 95d 10.244.0.4 kind-control-plane <none> <none>
NAMESPACE NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE SELECTOR
default service/apple-service ClusterIP 10.96.118.150 <none> 5678/TCP 3m5s app=apple
default service/banana-service ClusterIP 10.96.155.84 <none> 5678/TCP 3m5s app=banana
default service/kubernetes ClusterIP 10.96.0.1 <none> 443/TCP 95d <none>
kube-system service/ic-ingress-nginx-notcontroller LoadBalancer 10.96.187.204 <pending> 80:30496/TCP,443:31975/TCP 3m14s app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
kube-system service/ic-ingress-nginx-notcontroller-admission ClusterIP 10.96.24.137 <none> 443/TCP 3m14s app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
kube-system service/kube-dns ClusterIP 10.96.0.10 <none> 53/UDP,53/TCP,9153/TCP 95d k8s-app=kube-dns
NAMESPACE NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE CONTAINERS IMAGES SELECTOR
kube-system daemonset.apps/kindnet 1 1 1 1 1 <none> 95d kindnet-cni docker.io/kindest/kindnetd:v20210326-1e038dc5 app=kindnet
kube-system daemonset.apps/kube-proxy 1 1 1 1 1 kubernetes.io/os=linux 95d kube-proxy k8s.gcr.io/kube-proxy:v1.21.1 k8s-app=kube-proxy
NAMESPACE NAME READY UP-TO-DATE AVAILABLE AGE CONTAINERS IMAGES SELECTOR
kube-system deployment.apps/coredns 2/2 2 2 95d coredns k8s.gcr.io/coredns/coredns:v1.8.0 k8s-app=kube-dns
kube-system deployment.apps/ic-ingress-nginx-notcontroller 1/1 1 1 3m14s controller k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
local-path-storage deployment.apps/local-path-provisioner 1/1 1 1 95d local-path-provisioner docker.io/rancher/local-path-provisioner:v0.0.14 app=local-path-provisioner
NAMESPACE NAME DESIRED CURRENT READY AGE CONTAINERS IMAGES SELECTOR
kube-system replicaset.apps/coredns-558bd4d5db 2 2 2 95d coredns k8s.gcr.io/coredns/coredns:v1.8.0 k8s-app=kube-dns,pod-template-hash=558bd4d5db
kube-system replicaset.apps/ic-ingress-nginx-notcontroller-67f9445d98 1 1 1 3m14s controller k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx,pod-template-hash=67f9445d98
local-path-storage replicaset.apps/local-path-provisioner-547f784dff 1 1 1 95d local-path-provisioner docker.io/rancher/local-path-provisioner:v0.0.14 app=local-path-provisioner,pod-template-hash=547f784dff
```
- `kubectl -n <ingresscontrollernamespace> describe po <ingresscontrollerpodname>`
```
kubectl -n kube-system describe po ic-ingress-nginx-notcontroller-67f9445d98-8tgm8
Name: ic-ingress-nginx-notcontroller-67f9445d98-8tgm8
Namespace: kube-system
Priority: 0
Node: kind-control-plane/172.18.0.2
Start Time: Wed, 01 Jun 2022 16:35:40 -0300
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/name=ingress-nginx
pod-template-hash=67f9445d98
Annotations: <none>
Status: Running
IP: 10.244.0.10
IPs:
IP: 10.244.0.10
Controlled By: ReplicaSet/ic-ingress-nginx-notcontroller-67f9445d98
Containers:
controller:
Container ID: containerd://5a4c0f5ed21200b6b82e0afebd175076056100a5235f92f8c0287d0055956c7e
Image: k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de
Image ID: k8s.gcr.io/ingress-nginx/controller@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de
Ports: 80/TCP, 443/TCP, 8443/TCP
Host Ports: 0/TCP, 0/TCP, 0/TCP
Args:
/nginx-ingress-controller
--publish-service=$(POD_NAMESPACE)/ic-ingress-nginx-notcontroller
--election-id=ingress-controller-leader
--controller-class=k8s.io/ingress-nginx
--configmap=$(POD_NAMESPACE)/ic-ingress-nginx-notcontroller
--validating-webhook=:8443
--validating-webhook-certificate=/usr/local/certificates/cert
--validating-webhook-key=/usr/local/certificates/key
State: Running
Started: Wed, 01 Jun 2022 16:35:41 -0300
Ready: True
Restart Count: 0
Requests:
cpu: 100m
memory: 90Mi
Liveness: http-get http://:10254/healthz delay=10s timeout=1s period=10s #success=1 #failure=5
Readiness: http-get http://:10254/healthz delay=10s timeout=1s period=10s #success=1 #failure=3
Environment:
POD_NAME: ic-ingress-nginx-notcontroller-67f9445d98-8tgm8 (v1:metadata.name)
POD_NAMESPACE: kube-system (v1:metadata.namespace)
LD_PRELOAD: /usr/local/lib/libmimalloc.so
Mounts:
/usr/local/certificates/ from webhook-cert (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4pcbr (ro)
Conditions:
Type Status
Initialized True
Ready True
ContainersReady True
PodScheduled True
Volumes:
webhook-cert:
Type: Secret (a volume populated by a Secret)
SecretName: ic-ingress-nginx-admission
Optional: false
kube-api-access-4pcbr:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
QoS Class: Burstable
Node-Selectors: kubernetes.io/os=linux
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 4m18s default-scheduler Successfully assigned kube-system/ic-ingress-nginx-notcontroller-67f9445d98-8tgm8 to kind-control-plane
Normal Pulled 4m18s kubelet Container image "k8s.gcr.io/ingress-nginx/controller:v1.1.1@sha256:0bc88eb15f9e7f84e8e56c14fa5735aaa488b840983f87bd79b1054190e660de" already present on machine
Normal Created 4m18s kubelet Created container controller
Normal Started 4m18s kubelet Started container controller
Normal RELOAD 4m16s nginx-ingress-controller NGINX reload triggered due to a change in configuration
```
- `kubectl -n <ingresscontrollernamespace> describe svc <ingresscontrollerservicename>`
```
kubectl -n kube-system describe svc ic-ingress-nginx-notcontroller
Name: ic-ingress-nginx-notcontroller
Namespace: kube-system
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=ingress-nginx
app.kubernetes.io/version=1.1.1
helm.sh/chart=ingress-nginx-4.0.15
Annotations: meta.helm.sh/release-name: ic-ingress-nginx
meta.helm.sh/release-namespace: kube-system
Selector: app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
Type: LoadBalancer
IP Family Policy: SingleStack
IP Families: IPv4
IP: 10.96.187.204
IPs: 10.96.187.204
Port: http 80/TCP
TargetPort: http/TCP
NodePort: http 30496/TCP
Endpoints: 10.244.0.10:80
Port: https 443/TCP
TargetPort: https/TCP
NodePort: https 31975/TCP
Endpoints: 10.244.0.10:443
Session Affinity: None
External Traffic Policy: Cluster
Events: <none>
kubectl -n kube-system describe svc ic-ingress-nginx-notcontroller-admission
Name: ic-ingress-nginx-notcontroller-admission
Namespace: kube-system
Labels: app.kubernetes.io/component=controller
app.kubernetes.io/instance=ic-ingress-nginx
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=ingress-nginx
app.kubernetes.io/version=1.1.1
helm.sh/chart=ingress-nginx-4.0.15
Annotations: meta.helm.sh/release-name: ic-ingress-nginx
meta.helm.sh/release-namespace: kube-system
Selector: app.kubernetes.io/component=controller,app.kubernetes.io/instance=ic-ingress-nginx,app.kubernetes.io/name=ingress-nginx
Type: ClusterIP
IP Family Policy: SingleStack
IP Families: IPv4
IP: 10.96.24.137
IPs: 10.96.24.137
Port: https-webhook 443/TCP
TargetPort: webhook/TCP
Endpoints: 10.244.0.10:8443
Session Affinity: None
Events: <none>
```
- **Current state of ingress object, if applicable**:
- `kubectl -n <appnnamespace> get all,ing -o wide`
```
kubectl get all,ing -o wide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
pod/apple-app 1/1 Running 0 7m7s 10.244.0.12 kind-control-plane <none> <none>
pod/banana-app 1/1 Running 0 7m7s 10.244.0.13 kind-control-plane <none> <none>
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE SELECTOR
service/apple-service ClusterIP 10.96.118.150 <none> 5678/TCP 7m7s app=apple
service/banana-service ClusterIP 10.96.155.84 <none> 5678/TCP 7m7s app=banana
service/kubernetes ClusterIP 10.96.0.1 <none> 443/TCP 95d <none>
```
- Snippet used to apply this app and ingress resource:
```
cat <<EOF> simple-ingress.yml
kind: Pod
apiVersion: v1
metadata:
name: apple-app
labels:
app: apple
spec:
containers:
- name: apple-app
image: hashicorp/http-echo
args:
- "-text=apple"
---
kind: Service
apiVersion: v1
metadata:
name: apple-service
spec:
selector:
app: apple
ports:
- port: 5678 # Default port for image
---
kind: Pod
apiVersion: v1
metadata:
name: banana-app
labels:
app: banana
spec:
containers:
- name: banana-app
image: hashicorp/http-echo
args:
- "-text=banana"
---
kind: Service
apiVersion: v1
metadata:
name: banana-service
spec:
selector:
app: banana
ports:
- port: 5678 # Default port for image
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: example-ingress
annotations:
ingress.kubernetes.io/rewrite-target: /
spec:
ingressClassName: nginx
rules:
- http:
paths:
- backend:
service:
name: apple-service
port:
number: 5678
path: /apple
pathType: ImplementationSpecific
- backend:
service:
name: banana-service
port:
number: 5678
path: /banana
pathType: ImplementationSpecific
EOF
kubectl apply -f simple-ingress.yml
```
**What happened**:
It seems like .Values.controller.name cannot be different than 'controller'. Changing it to someting else results in an certificate not valid for domain error:
```
kubectl apply -f simple-ingress.yml
pod/apple-app unchanged
service/apple-service unchanged
pod/banana-app unchanged
service/banana-service unchanged
Error from server (InternalError): error when creating "simple-ingress.yml": Internal error occurred: failed calling webhook "validate.nginx.ingress.kubernetes.io": Post "https://ic-ingress-nginx-notcontroller-admission.kube-system.svc:443/networking/v1/ingresses?timeout=10s": x509: certificate is valid for ic-ingress-nginx-controller-admission, ic-ingress-nginx-controller-admission.kube-system.svc, not ic-ingress-nginx-notcontroller-admission.kube-system.svc
```
**What you expected to happen**:
The Ingress resource should be created without domain errors.
**How to reproduce it**:
* create the values file
```
cat <<EOF> values.yml
controller:
service:
ipFamilies:
- IPv4
EOF
```
* install version 4.0.15 using the previously created values.yml file
```
helm install ic-ingress-nginx ingress-nginx/ingress-nginx --version 4.0.15 -n kube-system -f values.yml
```
* create the ingress resource ( sucessfully this time )
```
cat <<EOF> simple-ingress.yml
kind: Pod
apiVersion: v1
metadata:
name: apple-app
labels:
app: apple
spec:
containers:
- name: apple-app
image: hashicorp/http-echo
args:
- "-text=apple"
---
kind: Service
apiVersion: v1
metadata:
name: apple-service
spec:
selector:
app: apple
ports:
- port: 5678 # Default port for image
---
kind: Pod
apiVersion: v1
metadata:
name: banana-app
labels:
app: banana
spec:
containers:
- name: banana-app
image: hashicorp/http-echo
args:
- "-text=banana"
---
kind: Service
apiVersion: v1
metadata:
name: banana-service
spec:
selector:
app: banana
ports:
- port: 5678 # Default port for image
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: example-ingress
annotations:
ingress.kubernetes.io/rewrite-target: /
spec:
ingressClassName: nginx
rules:
- http:
paths:
- backend:
service:
name: apple-service
port:
number: 5678
path: /apple
pathType: ImplementationSpecific
- backend:
service:
name: banana-service
port:
number: 5678
path: /banana
pathType: ImplementationSpecific
EOF
kubectl apply -f simple-ingress.yml
```
* Check that it works
```
POD_NAME=$(kubectl get pods -n kube-system -l app.kubernetes.io/name=ingress-nginx -o NAME)
kubectl exec -it -n kube-system $POD_NAME -- curl -H 'Host: foo.bar' localhost/banana
```
* Edit values.yml file and add an override for .Values.controller.name to something other than controller
```
cat <<EOF> values.yml
controller:
name: notcontroller
service:
ipFamilies:
- IPv4
EOF
```
* Run helm update using that upgrade values.yml file
```
helm upgrade ic-ingress-nginx ingress-nginx/ingress-nginx --version 4.0.15 -n kube-system -f values.yml
```
* Delete the resources of simple-ingress.yml
```
kubectl delete -f simple-ingress.yml
```
* Create the resources defined in simple-ingress.yml file again ( This time the ERROR happens )
```
kubectl apply -f simple-ingress.yml
pod/apple-app created
service/apple-service created
pod/banana-app created
service/banana-service created
Error from server (InternalError): error when creating "simple-ingress.yml": Internal error occurred: failed calling webhook "validate.nginx.ingress.kubernetes.io": Post "https://ic-ingress-nginx-notcontroller-admission.kube-system.svc:443/networking/v1/ingresses?timeout=10s": x509: certificate is valid for ic-ingress-nginx-controller-admission, ic-ingress-nginx-controller-admission.kube-system.svc, not ic-ingress-nginx-notcontroller-admission.kube-system.svc
```
<!---
As minimally and precisely as possible. Keep in mind we do not have access to your cluster or application.
Help up us (if possible) reproducing the issue using minikube or kind.
## Install minikube/kind
- Minikube https://minikube.sigs.k8s.io/docs/start/
- Kind https://kind.sigs.k8s.io/docs/user/quick-start/
## Install the ingress controller
kubectl apply -f https://raw.githubusercontent.com/kubernetes/ingress-nginx/main/deploy/static/provider/baremetal/deploy.yaml
## Install an application that will act as default backend (is just an echo app)
kubectl apply -f https://raw.githubusercontent.com/kubernetes/ingress-nginx/main/docs/examples/http-svc.yaml
## Create an ingress (please add any additional annotation required)
echo "
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: foo-bar
annotations:
kubernetes.io/ingress.class: nginx
spec:
ingressClassName: nginx # omit this if you're on controller version below 1.0.0
rules:
- host: foo.bar
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: http-svc
port:
number: 80
" | kubectl apply -f -
## make a request
POD_NAME=$(k get pods -n ingress-nginx -l app.kubernetes.io/name=ingress-nginx -o NAME)
kubectl exec -it -n ingress-nginx $POD_NAME -- curl -H 'Host: foo.bar' localhost
--->
<!-- If this is actually about documentation, uncomment the following block -->
<!--
/kind documentation
/remove-kind bug
-->
| priority | changing values controller name results in a certificate not valid for domain error welcome to ingress nginx for a smooth issue process try to answer the following questions don t worry if they re not all applicable just try to include what you can if you need to include code snippets or logs please put them in fenced code blocks if they re super long please use the details tag like super long log lots of stuff important please complete the next sections or the issue will be closed this questions are the first thing we need to know to understand the context nginx ingress controller version exec into the pod and run nginx ingress controller version pod namespace ingress nginx pod name kubectl get pods n pod namespace l app kubernetes io name ingress nginx field selector status phase running o jsonpath items metadata name kubectl exec it pod name n pod namespace nginx ingress controller version kubernetes version use kubectl version environment cloud provider or hardware configuration oracle virtual box os e g from etc os release name ubuntu version lts focal fossa id ubuntu id like debian pretty name ubuntu lts version id home url support url bug report url privacy policy url version codename focal ubuntu codename focal kernel e g uname a linux alienware generic ubuntu smp wed may utc gnu linux install tools seen this issue in both a kubeadm and minukube created clusters basic cluster related info kubectl version warning this version information is deprecated and will be replaced with the output from kubectl version short use output yaml json to get the full version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux kustomize version server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux warning version difference between client and server kubectl get nodes o wide kubectl get nodes o wide name status roles age version internal ip external ip os image kernel version container runtime kind control plane ready control plane master ubuntu generic containerd how was the ingress nginx controller installed if helm was used then please show output of helm ls a grep i ingress helm ls a grep i ingress ic ingress nginx kube system deployed ingress nginx if helm was used then please show output of helm n get values helm n kube system get values ic ingress nginx user supplied values controller name notcontroller service ipfamilies current state of the controller kubectl describe ingressclasses kubectl describe ingressclasses name nginx labels app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io managed by helm app kubernetes io name ingress nginx app kubernetes io version helm sh chart ingress nginx annotations meta helm sh release name ic ingress nginx meta helm sh release namespace kube system controller io ingress nginx events kubectl n get all a o wide kubectl n kube system get all a o wide namespace name ready status restarts age ip node nominated node readiness gates default pod apple app running kind control plane default pod banana app running kind control plane kube system pod coredns running kind control plane kube system pod coredns xxsfx running kind control plane kube system pod etcd kind control plane running kind control plane kube system pod ic ingress nginx notcontroller running kind control plane kube system pod kindnet running kind control plane kube system pod kube apiserver kind control plane running kind control plane kube system pod kube controller manager kind control plane running kind control plane kube system pod kube proxy running kind control plane kube system pod kube scheduler kind control plane running kind control plane local path storage pod local path provisioner running kind control plane namespace name type cluster ip external ip port s age selector default service apple service clusterip tcp app apple default service banana service clusterip tcp app banana default service kubernetes clusterip tcp kube system service ic ingress nginx notcontroller loadbalancer tcp tcp app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx kube system service ic ingress nginx notcontroller admission clusterip tcp app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx kube system service kube dns clusterip udp tcp tcp app kube dns namespace name desired current ready up to date available node selector age containers images selector kube system daemonset apps kindnet kindnet cni docker io kindest kindnetd app kindnet kube system daemonset apps kube proxy kubernetes io os linux kube proxy gcr io kube proxy app kube proxy namespace name ready up to date available age containers images selector kube system deployment apps coredns coredns gcr io coredns coredns app kube dns kube system deployment apps ic ingress nginx notcontroller controller gcr io ingress nginx controller app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx local path storage deployment apps local path provisioner local path provisioner docker io rancher local path provisioner app local path provisioner namespace name desired current ready age containers images selector kube system replicaset apps coredns coredns gcr io coredns coredns app kube dns pod template hash kube system replicaset apps ic ingress nginx notcontroller controller gcr io ingress nginx controller app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx pod template hash local path storage replicaset apps local path provisioner local path provisioner docker io rancher local path provisioner app local path provisioner pod template hash kubectl n describe po kubectl n kube system describe po ic ingress nginx notcontroller name ic ingress nginx notcontroller namespace kube system priority node kind control plane start time wed jun labels app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx pod template hash annotations status running ip ips ip controlled by replicaset ic ingress nginx notcontroller containers controller container id containerd image gcr io ingress nginx controller image id gcr io ingress nginx controller ports tcp tcp tcp host ports tcp tcp tcp args nginx ingress controller publish service pod namespace ic ingress nginx notcontroller election id ingress controller leader controller class io ingress nginx configmap pod namespace ic ingress nginx notcontroller validating webhook validating webhook certificate usr local certificates cert validating webhook key usr local certificates key state running started wed jun ready true restart count requests cpu memory liveness http get delay timeout period success failure readiness http get delay timeout period success failure environment pod name ic ingress nginx notcontroller metadata name pod namespace kube system metadata namespace ld preload usr local lib libmimalloc so mounts usr local certificates from webhook cert ro var run secrets kubernetes io serviceaccount from kube api access ro conditions type status initialized true ready true containersready true podscheduled true volumes webhook cert type secret a volume populated by a secret secretname ic ingress nginx admission optional false kube api access type projected a volume that contains injected data from multiple sources tokenexpirationseconds configmapname kube root ca crt configmapoptional downwardapi true qos class burstable node selectors kubernetes io os linux tolerations node kubernetes io not ready noexecute op exists for node kubernetes io unreachable noexecute op exists for events type reason age from message normal scheduled default scheduler successfully assigned kube system ic ingress nginx notcontroller to kind control plane normal pulled kubelet container image gcr io ingress nginx controller already present on machine normal created kubelet created container controller normal started kubelet started container controller normal reload nginx ingress controller nginx reload triggered due to a change in configuration kubectl n describe svc kubectl n kube system describe svc ic ingress nginx notcontroller name ic ingress nginx notcontroller namespace kube system labels app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io managed by helm app kubernetes io name ingress nginx app kubernetes io version helm sh chart ingress nginx annotations meta helm sh release name ic ingress nginx meta helm sh release namespace kube system selector app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx type loadbalancer ip family policy singlestack ip families ip ips port http tcp targetport http tcp nodeport http tcp endpoints port https tcp targetport https tcp nodeport https tcp endpoints session affinity none external traffic policy cluster events kubectl n kube system describe svc ic ingress nginx notcontroller admission name ic ingress nginx notcontroller admission namespace kube system labels app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io managed by helm app kubernetes io name ingress nginx app kubernetes io version helm sh chart ingress nginx annotations meta helm sh release name ic ingress nginx meta helm sh release namespace kube system selector app kubernetes io component controller app kubernetes io instance ic ingress nginx app kubernetes io name ingress nginx type clusterip ip family policy singlestack ip families ip ips port https webhook tcp targetport webhook tcp endpoints session affinity none events current state of ingress object if applicable kubectl n get all ing o wide kubectl get all ing o wide name ready status restarts age ip node nominated node readiness gates pod apple app running kind control plane pod banana app running kind control plane name type cluster ip external ip port s age selector service apple service clusterip tcp app apple service banana service clusterip tcp app banana service kubernetes clusterip tcp snippet used to apply this app and ingress resource cat simple ingress yml kind pod apiversion metadata name apple app labels app apple spec containers name apple app image hashicorp http echo args text apple kind service apiversion metadata name apple service spec selector app apple ports port default port for image kind pod apiversion metadata name banana app labels app banana spec containers name banana app image hashicorp http echo args text banana kind service apiversion metadata name banana service spec selector app banana ports port default port for image apiversion networking io kind ingress metadata name example ingress annotations ingress kubernetes io rewrite target spec ingressclassname nginx rules http paths backend service name apple service port number path apple pathtype implementationspecific backend service name banana service port number path banana pathtype implementationspecific eof kubectl apply f simple ingress yml what happened it seems like values controller name cannot be different than controller changing it to someting else results in an certificate not valid for domain error kubectl apply f simple ingress yml pod apple app unchanged service apple service unchanged pod banana app unchanged service banana service unchanged error from server internalerror error when creating simple ingress yml internal error occurred failed calling webhook validate nginx ingress kubernetes io post certificate is valid for ic ingress nginx controller admission ic ingress nginx controller admission kube system svc not ic ingress nginx notcontroller admission kube system svc what you expected to happen the ingress resource should be created without domain errors how to reproduce it create the values file cat values yml controller service ipfamilies eof install version using the previously created values yml file helm install ic ingress nginx ingress nginx ingress nginx version n kube system f values yml create the ingress resource sucessfully this time cat simple ingress yml kind pod apiversion metadata name apple app labels app apple spec containers name apple app image hashicorp http echo args text apple kind service apiversion metadata name apple service spec selector app apple ports port default port for image kind pod apiversion metadata name banana app labels app banana spec containers name banana app image hashicorp http echo args text banana kind service apiversion metadata name banana service spec selector app banana ports port default port for image apiversion networking io kind ingress metadata name example ingress annotations ingress kubernetes io rewrite target spec ingressclassname nginx rules http paths backend service name apple service port number path apple pathtype implementationspecific backend service name banana service port number path banana pathtype implementationspecific eof kubectl apply f simple ingress yml check that it works pod name kubectl get pods n kube system l app kubernetes io name ingress nginx o name kubectl exec it n kube system pod name curl h host foo bar localhost banana edit values yml file and add an override for values controller name to something other than controller cat values yml controller name notcontroller service ipfamilies eof run helm update using that upgrade values yml file helm upgrade ic ingress nginx ingress nginx ingress nginx version n kube system f values yml delete the resources of simple ingress yml kubectl delete f simple ingress yml create the resources defined in simple ingress yml file again this time the error happens kubectl apply f simple ingress yml pod apple app created service apple service created pod banana app created service banana service created error from server internalerror error when creating simple ingress yml internal error occurred failed calling webhook validate nginx ingress kubernetes io post certificate is valid for ic ingress nginx controller admission ic ingress nginx controller admission kube system svc not ic ingress nginx notcontroller admission kube system svc as minimally and precisely as possible keep in mind we do not have access to your cluster or application help up us if possible reproducing the issue using minikube or kind install minikube kind minikube kind install the ingress controller kubectl apply f install an application that will act as default backend is just an echo app kubectl apply f create an ingress please add any additional annotation required echo apiversion networking io kind ingress metadata name foo bar annotations kubernetes io ingress class nginx spec ingressclassname nginx omit this if you re on controller version below rules host foo bar http paths path pathtype prefix backend service name http svc port number kubectl apply f make a request pod name k get pods n ingress nginx l app kubernetes io name ingress nginx o name kubectl exec it n ingress nginx pod name curl h host foo bar localhost kind documentation remove kind bug | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.