Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
142,122
13,016,726,320
IssuesEvent
2020-07-26 08:22:14
anitab-org/anitab-org.github.io
https://api.github.com/repos/anitab-org/anitab-org.github.io
closed
Improvement Of README.md File.
Category: Documentation/Training First Timers Only
### Description As a developer, While setting up the remote upstream, documentation was unclear. The README.md file describes the following. "When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command - git remote add upstream https://github.com/anitab-org/anitab-org.github.io." ### Expected Result "When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command - **cd anitab-org.github.io** git remote add upstream https://github.com/anitab-org/anitab-org.github.io" ### Mocks ![Screenshot (241)_LI](https://user-images.githubusercontent.com/43727167/87247209-fa6bbc80-c46f-11ea-81ca-912658253fbd.jpg)
1.0
Improvement Of README.md File. - ### Description As a developer, While setting up the remote upstream, documentation was unclear. The README.md file describes the following. "When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command - git remote add upstream https://github.com/anitab-org/anitab-org.github.io." ### Expected Result "When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command - **cd anitab-org.github.io** git remote add upstream https://github.com/anitab-org/anitab-org.github.io" ### Mocks ![Screenshot (241)_LI](https://user-images.githubusercontent.com/43727167/87247209-fa6bbc80-c46f-11ea-81ca-912658253fbd.jpg)
non_process
improvement of readme md file description as a developer while setting up the remote upstream documentation was unclear the readme md file describes the following when a repository is cloned it has a default remote named origin that points to your fork on github not the original repository it was forked from to keep track of the original repository you should add another remote named upstream for this project it can be done by running the following command git remote add upstream expected result when a repository is cloned it has a default remote named origin that points to your fork on github not the original repository it was forked from to keep track of the original repository you should add another remote named upstream for this project it can be done by running the following command cd anitab org github io git remote add upstream mocks
0
13,376
15,837,949,036
IssuesEvent
2021-04-06 21:35:06
googleapis/python-pubsub
https://api.github.com/repos/googleapis/python-pubsub
closed
Unit tests must not use/expect credentials from the environment
api: pubsub type: process
```bash $ env | grep GOOGLE && echo YES || echo NO NO $ git log -1 commit 469ebaa3c449c881089dfc657da5902c1d031803 (HEAD -> master, origin/master, origin/HEAD) Author: Peter Lamut <plamut@users.noreply.github.com> Date: Fri Apr 2 09:26:10 2021 +0200 chore: regenerate GAPIC layer with latest changes (#345) $ nox -e unit-3.8 nox > Running session unit-3.8 nox > Creating virtual environment (virtualenv) using python3.8 in .nox/unit-3-8 nox > pip install asyncmock pytest-asyncio nox > pip install mock pytest pytest-cov nox > pip install -e . nox > py.test --quiet --junitxml=unit_3.8_sponge_log.xml --cov=google/cloud --cov=tests/unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit ........................................................................ [ 8%] ........................................................................ [ 17%] ........................................................................ [ 26%] ........................................................................ [ 34%] ........................................................................ [ 43%] ........................................................................ [ 52%] ........................................................................ [ 61%] ........................................................................ [ 69%] ........................................................................ [ 78%] ........................................................................ [ 87%] ........................................................................ [ 96%] ............................... [100%] =============================== warnings summary =============================== ... tests/unit/pubsub_v1/publisher/test_publisher_client.py: 4 warnings tests/unit/pubsub_v1/subscriber/test_subscriber_client.py: 9 warnings /home/tseaver/projects/agendaless/Google/src/python-pubsub/.nox/unit-3-8/lib/python3.8/site-packages/google/auth/_default.py:70: UserWarning: Your application has authenticated using end user credentials from Google Cloud SDK without a quota project. You might receive a "quota exceeded" or "API not enabled" error. We recommend you rerun `gcloud auth application-default login` and make sure a quota project is added. Or you can use service accounts instead. For more information about service accounts, see https://cloud.google.com/docs/authentication/ warnings.warn(_CLOUD_SDK_CREDENTIALS_WARNING) -- Docs: https://docs.pytest.org/en/stable/warnings.html - generated xml file: /home/tseaver/projects/agendaless/Google/src/python-pubsub/unit_3.8_sponge_log.xml - 823 passed, 15 warnings in 13.94s ``` Unit tests should always pass explicit dummy credentials (e.g., see [`test_init`](https://github.com/googleapis/python-pubsub/blob/469ebaa3c449c881089dfc657da5902c1d031803/tests/unit/pubsub_v1/publisher/test_publisher_client.py#L54-L56)).
1.0
Unit tests must not use/expect credentials from the environment - ```bash $ env | grep GOOGLE && echo YES || echo NO NO $ git log -1 commit 469ebaa3c449c881089dfc657da5902c1d031803 (HEAD -> master, origin/master, origin/HEAD) Author: Peter Lamut <plamut@users.noreply.github.com> Date: Fri Apr 2 09:26:10 2021 +0200 chore: regenerate GAPIC layer with latest changes (#345) $ nox -e unit-3.8 nox > Running session unit-3.8 nox > Creating virtual environment (virtualenv) using python3.8 in .nox/unit-3-8 nox > pip install asyncmock pytest-asyncio nox > pip install mock pytest pytest-cov nox > pip install -e . nox > py.test --quiet --junitxml=unit_3.8_sponge_log.xml --cov=google/cloud --cov=tests/unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit ........................................................................ [ 8%] ........................................................................ [ 17%] ........................................................................ [ 26%] ........................................................................ [ 34%] ........................................................................ [ 43%] ........................................................................ [ 52%] ........................................................................ [ 61%] ........................................................................ [ 69%] ........................................................................ [ 78%] ........................................................................ [ 87%] ........................................................................ [ 96%] ............................... [100%] =============================== warnings summary =============================== ... tests/unit/pubsub_v1/publisher/test_publisher_client.py: 4 warnings tests/unit/pubsub_v1/subscriber/test_subscriber_client.py: 9 warnings /home/tseaver/projects/agendaless/Google/src/python-pubsub/.nox/unit-3-8/lib/python3.8/site-packages/google/auth/_default.py:70: UserWarning: Your application has authenticated using end user credentials from Google Cloud SDK without a quota project. You might receive a "quota exceeded" or "API not enabled" error. We recommend you rerun `gcloud auth application-default login` and make sure a quota project is added. Or you can use service accounts instead. For more information about service accounts, see https://cloud.google.com/docs/authentication/ warnings.warn(_CLOUD_SDK_CREDENTIALS_WARNING) -- Docs: https://docs.pytest.org/en/stable/warnings.html - generated xml file: /home/tseaver/projects/agendaless/Google/src/python-pubsub/unit_3.8_sponge_log.xml - 823 passed, 15 warnings in 13.94s ``` Unit tests should always pass explicit dummy credentials (e.g., see [`test_init`](https://github.com/googleapis/python-pubsub/blob/469ebaa3c449c881089dfc657da5902c1d031803/tests/unit/pubsub_v1/publisher/test_publisher_client.py#L54-L56)).
process
unit tests must not use expect credentials from the environment bash env grep google echo yes echo no no git log commit head master origin master origin head author peter lamut date fri apr chore regenerate gapic layer with latest changes nox e unit nox running session unit nox creating virtual environment virtualenv using in nox unit nox pip install asyncmock pytest asyncio nox pip install mock pytest pytest cov nox pip install e nox py test quiet junitxml unit sponge log xml cov google cloud cov tests unit cov append cov config coveragerc cov report cov fail under tests unit warnings summary tests unit pubsub publisher test publisher client py warnings tests unit pubsub subscriber test subscriber client py warnings home tseaver projects agendaless google src python pubsub nox unit lib site packages google auth default py userwarning your application has authenticated using end user credentials from google cloud sdk without a quota project you might receive a quota exceeded or api not enabled error we recommend you rerun gcloud auth application default login and make sure a quota project is added or you can use service accounts instead for more information about service accounts see warnings warn cloud sdk credentials warning docs generated xml file home tseaver projects agendaless google src python pubsub unit sponge log xml passed warnings in unit tests should always pass explicit dummy credentials e g see
1
621,349
19,583,498,279
IssuesEvent
2022-01-05 01:49:08
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.donateblood.com.au - Design is broken
browser-firefox priority-normal severity-critical engine-gecko
<!-- @browser: Firefox 90.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:90.0) Gecko/20100101 Firefox/90.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/81122 --> **URL**: https://www.donateblood.com.au/ **Browser / Version**: Firefox 90.0 **Operating System**: Mac OS X 10.15 **Tested Another Browser**: Yes Safari **Problem type**: Design is broken **Description**: Items are overlapped **Steps to Reproduce**: I loaded the website by copying the URL https://www.donateblood.com.au/ into the URL bar in Firefox and hitting the ENTER key. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/7/2d88002f-203b-4de6-848c-dcdbaa651611.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.donateblood.com.au - Design is broken - <!-- @browser: Firefox 90.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:90.0) Gecko/20100101 Firefox/90.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/81122 --> **URL**: https://www.donateblood.com.au/ **Browser / Version**: Firefox 90.0 **Operating System**: Mac OS X 10.15 **Tested Another Browser**: Yes Safari **Problem type**: Design is broken **Description**: Items are overlapped **Steps to Reproduce**: I loaded the website by copying the URL https://www.donateblood.com.au/ into the URL bar in Firefox and hitting the ENTER key. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/7/2d88002f-203b-4de6-848c-dcdbaa651611.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
design is broken url browser version firefox operating system mac os x tested another browser yes safari problem type design is broken description items are overlapped steps to reproduce i loaded the website by copying the url into the url bar in firefox and hitting the enter key view the screenshot img alt screenshot src browser configuration none from with ❤️
0
9,718
12,716,601,222
IssuesEvent
2020-06-24 02:26:46
OUDcollective/twenty20times
https://api.github.com/repos/OUDcollective/twenty20times
opened
Understanding the GitHub flow · GitHub Guides
workflow-process
![Screen Shot from awesomescreenshot.com](https://www.awesomescreenshot.com/api/v1/destination/image/show?ImageKey=tm-3919-15894-777a198076e1d4a40b936aa0c2ef3b4a) ## Deploy With GitHub, you can deploy from a branch for final testing in production before merging to master. --- **Source URL**: [https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/) <table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
1.0
Understanding the GitHub flow · GitHub Guides - ![Screen Shot from awesomescreenshot.com](https://www.awesomescreenshot.com/api/v1/destination/image/show?ImageKey=tm-3919-15894-777a198076e1d4a40b936aa0c2ef3b4a) ## Deploy With GitHub, you can deploy from a branch for final testing in production before merging to master. --- **Source URL**: [https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/) <table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
process
understanding the github flow · github guides deploy with github you can deploy from a branch for final testing in production before merging to master source url browser chrome os windows bit screen size viewport size pixel ratio zoom level
1
11,660
14,525,321,069
IssuesEvent
2020-12-14 12:46:53
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Add de-dot processor that converts dotted field names to nested objects
:Processors Filebeat Team:Services enhancement
**Background** This is a requirement that came up in the https://github.com/elastic/ecs-logging initiative. In summary, we're trying to make logging simpler by logging ECS-compliant JSON to a file that Filebeat can just forward to Elasticsearch. However, due to readability, performance, and other technical reasons, it's not always possible for the loggers to produce a correctly nested JSON structure. Some fields, like `log.logger` may be represented via a field name containing a dot (`"log.logger": "INFO"`) while others may be nested (`"foo": { "bar": "baz"}`). More context here: https://github.com/elastic/ecs-logging-java/issues/51 **The problem** When processing log data, for example with an Elasticsearch Ingest pipeline, we need all fields to be nested. Otherwise, the user doesn't know whether to access a field via `doc["foo.bar"]` or via `doc["foo"]["bar"]`. We don't want users to have knowledge about which fields are nested vs dotted as this is an implementation detail that can vary with different `ecs-logging` implementations and may even change for the same implementation. Also, ECS defines that fields should be always nested. **Describe the enhancement:** We'd like to have a Filebeat processor that expands all dotted field names to nested objects. This would decouple the representation in the log file from how the documents are supposed to look once they hit the ingest node processing pipeline. **Concerns** - Performance: This might be a performance hit but I suspect other processors, like grok, to be much more processing intensive. - If the JSON is already fully nested, we could short-circuit the processing - We could require that once we're in a nested context, dots are no longer replaced with dotting. - Allowed: `"foo.bar": {"baz": "qux"}` - Disallowed: `"foo": {"bar.baz": "qux"}` - Given this restricion, the de-dotting can be done very efficiently by sorting the keys alphabetically and processing the JSON similar to how a SAX-parser works (@urso's idea). **Open Questions** What to do when conflicts occur. - Incompatible mappings: `"foo.bar": "baz"`, `"foo": "bar"` (foo is both an object and a string). - Duplicate keys: `"foo.bar": "baz"`, `"foo.bar": "qux"`
1.0
Add de-dot processor that converts dotted field names to nested objects - **Background** This is a requirement that came up in the https://github.com/elastic/ecs-logging initiative. In summary, we're trying to make logging simpler by logging ECS-compliant JSON to a file that Filebeat can just forward to Elasticsearch. However, due to readability, performance, and other technical reasons, it's not always possible for the loggers to produce a correctly nested JSON structure. Some fields, like `log.logger` may be represented via a field name containing a dot (`"log.logger": "INFO"`) while others may be nested (`"foo": { "bar": "baz"}`). More context here: https://github.com/elastic/ecs-logging-java/issues/51 **The problem** When processing log data, for example with an Elasticsearch Ingest pipeline, we need all fields to be nested. Otherwise, the user doesn't know whether to access a field via `doc["foo.bar"]` or via `doc["foo"]["bar"]`. We don't want users to have knowledge about which fields are nested vs dotted as this is an implementation detail that can vary with different `ecs-logging` implementations and may even change for the same implementation. Also, ECS defines that fields should be always nested. **Describe the enhancement:** We'd like to have a Filebeat processor that expands all dotted field names to nested objects. This would decouple the representation in the log file from how the documents are supposed to look once they hit the ingest node processing pipeline. **Concerns** - Performance: This might be a performance hit but I suspect other processors, like grok, to be much more processing intensive. - If the JSON is already fully nested, we could short-circuit the processing - We could require that once we're in a nested context, dots are no longer replaced with dotting. - Allowed: `"foo.bar": {"baz": "qux"}` - Disallowed: `"foo": {"bar.baz": "qux"}` - Given this restricion, the de-dotting can be done very efficiently by sorting the keys alphabetically and processing the JSON similar to how a SAX-parser works (@urso's idea). **Open Questions** What to do when conflicts occur. - Incompatible mappings: `"foo.bar": "baz"`, `"foo": "bar"` (foo is both an object and a string). - Duplicate keys: `"foo.bar": "baz"`, `"foo.bar": "qux"`
process
add de dot processor that converts dotted field names to nested objects background this is a requirement that came up in the initiative in summary we re trying to make logging simpler by logging ecs compliant json to a file that filebeat can just forward to elasticsearch however due to readability performance and other technical reasons it s not always possible for the loggers to produce a correctly nested json structure some fields like log logger may be represented via a field name containing a dot log logger info while others may be nested foo bar baz more context here the problem when processing log data for example with an elasticsearch ingest pipeline we need all fields to be nested otherwise the user doesn t know whether to access a field via doc or via doc we don t want users to have knowledge about which fields are nested vs dotted as this is an implementation detail that can vary with different ecs logging implementations and may even change for the same implementation also ecs defines that fields should be always nested describe the enhancement we d like to have a filebeat processor that expands all dotted field names to nested objects this would decouple the representation in the log file from how the documents are supposed to look once they hit the ingest node processing pipeline concerns performance this might be a performance hit but i suspect other processors like grok to be much more processing intensive if the json is already fully nested we could short circuit the processing we could require that once we re in a nested context dots are no longer replaced with dotting allowed foo bar baz qux disallowed foo bar baz qux given this restricion the de dotting can be done very efficiently by sorting the keys alphabetically and processing the json similar to how a sax parser works urso s idea open questions what to do when conflicts occur incompatible mappings foo bar baz foo bar foo is both an object and a string duplicate keys foo bar baz foo bar qux
1
10,999
13,788,690,448
IssuesEvent
2020-10-09 07:40:11
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Uncaught TypeError: Illegal invocation
SYSTEM: iframe processing TYPE: bug support center
The page is not loaded, there is the following error in DevTools console: ``` Uncaught TypeError: Illegal invocation at Window.addEventListener (hammerhead.js:7) at <anonymous>:2355:89162 at Object.<anonymous> (<anonymous>:2355:89829) at Object.jQuery (<anonymous>:2355:89881) at n (<anonymous>:15:228) at Object.<anonymous> (<anonymous>:2355:87883) at Object.jQuery (<anonymous>:2355:88125) at n (<anonymous>:15:228) at Object.jQuery (<anonymous>:2355:87042) at n (<anonymous>:15:228) ``` Please see the [T915177](https://supportcenter.devexpress.com/internal/ticket/details/T915177) private ticket for details. testcafe-hammerhead version: 17.1.11
1.0
Uncaught TypeError: Illegal invocation - The page is not loaded, there is the following error in DevTools console: ``` Uncaught TypeError: Illegal invocation at Window.addEventListener (hammerhead.js:7) at <anonymous>:2355:89162 at Object.<anonymous> (<anonymous>:2355:89829) at Object.jQuery (<anonymous>:2355:89881) at n (<anonymous>:15:228) at Object.<anonymous> (<anonymous>:2355:87883) at Object.jQuery (<anonymous>:2355:88125) at n (<anonymous>:15:228) at Object.jQuery (<anonymous>:2355:87042) at n (<anonymous>:15:228) ``` Please see the [T915177](https://supportcenter.devexpress.com/internal/ticket/details/T915177) private ticket for details. testcafe-hammerhead version: 17.1.11
process
uncaught typeerror illegal invocation the page is not loaded there is the following error in devtools console uncaught typeerror illegal invocation at window addeventlistener hammerhead js at at object at object jquery at n at object at object jquery at n at object jquery at n please see the private ticket for details testcafe hammerhead version
1
214,966
24,126,376,888
IssuesEvent
2022-09-21 01:04:29
dmartinez777/Tracking
https://api.github.com/repos/dmartinez777/Tracking
closed
WS-2021-0013 (Medium) detected in laravel/framework-v5.8.35 - autoclosed
security vulnerability
## WS-2021-0013 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.8.35</b></p></summary> <p>The Laravel Framework.</p> <p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197">https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197</a></p> <p> Dependency Hierarchy: - :x: **laravel/framework-v5.8.35** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Laravel is a web application framework. Versions of Laravel before 6.20.14, 7.30.4 and 8.24.0 contain a query binding exploitation. If a request is crafted where a field that is normally a non-array value is an array, and that input is not validated or cast to its expected type before being passed to the query builder, an unexpected number of query bindings can be added to the query. In some situations, this will simply lead to no results being returned by the query builder; however, it is possible certain queries could be affected in a way that causes the query to return unexpected results. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://github.com/laravel/framework/commit/2d9b970257bca7a176be897ec18dd5f6ffc5497f>WS-2021-0013</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-x7p5-p2c9-phvg">https://github.com/advisories/GHSA-x7p5-p2c9-phvg</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution: laravel/framework - 6.20.14, 7.30.4, 8.24.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0013 (Medium) detected in laravel/framework-v5.8.35 - autoclosed - ## WS-2021-0013 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.8.35</b></p></summary> <p>The Laravel Framework.</p> <p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197">https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197</a></p> <p> Dependency Hierarchy: - :x: **laravel/framework-v5.8.35** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Laravel is a web application framework. Versions of Laravel before 6.20.14, 7.30.4 and 8.24.0 contain a query binding exploitation. If a request is crafted where a field that is normally a non-array value is an array, and that input is not validated or cast to its expected type before being passed to the query builder, an unexpected number of query bindings can be added to the query. In some situations, this will simply lead to no results being returned by the query builder; however, it is possible certain queries could be affected in a way that causes the query to return unexpected results. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://github.com/laravel/framework/commit/2d9b970257bca7a176be897ec18dd5f6ffc5497f>WS-2021-0013</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-x7p5-p2c9-phvg">https://github.com/advisories/GHSA-x7p5-p2c9-phvg</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution: laravel/framework - 6.20.14, 7.30.4, 8.24.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in laravel framework autoclosed ws medium severity vulnerability vulnerable library laravel framework the laravel framework library home page a href dependency hierarchy x laravel framework vulnerable library vulnerability details laravel is a web application framework versions of laravel before and contain a query binding exploitation if a request is crafted where a field that is normally a non array value is an array and that input is not validated or cast to its expected type before being passed to the query builder an unexpected number of query bindings can be added to the query in some situations this will simply lead to no results being returned by the query builder however it is possible certain queries could be affected in a way that causes the query to return unexpected results publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope changed impact metrics confidentiality impact high integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution laravel framework step up your open source security game with mend
0
19,566
25,887,826,902
IssuesEvent
2022-12-14 15:43:29
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_cuda_simple (__main__.TestMultiprocessing)
module: multiprocessing module: cuda triaged module: flaky-tests skipped
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_cuda_simple&suite=TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8793519436). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_cuda_simple` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. cc @VitalyFedyunin @ngimel
1.0
DISABLED test_cuda_simple (__main__.TestMultiprocessing) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_cuda_simple&suite=TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8793519436). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_cuda_simple` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. cc @VitalyFedyunin @ngimel
process
disabled test cuda simple main testmultiprocessing platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test cuda simple there should be several instances run as flaky tests are rerun in ci from which you can study the logs cc vitalyfedyunin ngimel
1
60,140
3,120,782,434
IssuesEvent
2015-09-05 01:59:04
framingeinstein/issues-test
https://api.github.com/repos/framingeinstein/issues-test
opened
SRP-17: Create Custom Admin Test Functions
priority:normal resolution:in-progress type:enhancement
Hi Meghan, It looks like we can create some custom code to setup an admin-only payment method that will allow you to create test orders to send test emails. This would take up to an hour of billable work. If you want us to proceed, just let me know, and I'll get it in the que (but not for 1.1, at least as of now). Let me know if you have any questions, Thanks, Abe
1.0
SRP-17: Create Custom Admin Test Functions - Hi Meghan, It looks like we can create some custom code to setup an admin-only payment method that will allow you to create test orders to send test emails. This would take up to an hour of billable work. If you want us to proceed, just let me know, and I'll get it in the que (but not for 1.1, at least as of now). Let me know if you have any questions, Thanks, Abe
non_process
srp create custom admin test functions hi meghan it looks like we can create some custom code to setup an admin only payment method that will allow you to create test orders to send test emails this would take up to an hour of billable work if you want us to proceed just let me know and i ll get it in the que but not for at least as of now let me know if you have any questions thanks abe
0
89,138
3,790,062,917
IssuesEvent
2016-03-21 20:08:05
ReactiveX/rxjs
https://api.github.com/repos/ReactiveX/rxjs
opened
Mono-repo, many packages
priority: critical
We want to convert the repository into a single repository from which we publish many packages. This means a reorganization of the repository to something more suitable. Any and all ideas here are welcome. Especially those that come from prior art.
1.0
Mono-repo, many packages - We want to convert the repository into a single repository from which we publish many packages. This means a reorganization of the repository to something more suitable. Any and all ideas here are welcome. Especially those that come from prior art.
non_process
mono repo many packages we want to convert the repository into a single repository from which we publish many packages this means a reorganization of the repository to something more suitable any and all ideas here are welcome especially those that come from prior art
0
445,094
12,826,084,727
IssuesEvent
2020-07-06 15:57:43
eclipse/dirigible
https://api.github.com/repos/eclipse/dirigible
closed
[EDM] Projection Entity type to be introduced
component-ide efforts-low priority-medium templates usability web-ide
An Entity which belongs to an external model (file) and it is used only as a projection. No editing of any attribute or property is allowed in this case. No generation of any output artefact is expected for this type.
1.0
[EDM] Projection Entity type to be introduced - An Entity which belongs to an external model (file) and it is used only as a projection. No editing of any attribute or property is allowed in this case. No generation of any output artefact is expected for this type.
non_process
projection entity type to be introduced an entity which belongs to an external model file and it is used only as a projection no editing of any attribute or property is allowed in this case no generation of any output artefact is expected for this type
0
107,278
23,382,330,095
IssuesEvent
2022-08-11 10:38:58
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Test failure JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh
arch-arm64 os-mac-os-x GCStress area-CodeGen-coreclr blocking-clean-ci-optional
Run: [runtime-coreclr gcstress-extra 20220724.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1900569&view=ms.vss-test-web.build-test-results-tab&runId=49457008&paneView=debug&resultId=108429) Failed test: ``` coreclr OSX arm64 Checked gcstress0xc_zapdisable_heapverify1 @ OSX.1200.ARM64.Open - JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh ``` **Error message:** ``` [createdump] Invalid process id: task_for_pid(4529) FAILED (os/kern) failure (5) [createdump] This failure may be because createdump or the application is not properly signed and entitled. [createdump] Failure took 0ms /private/tmp/helix/working/A9B009AA/w/B9F709BA/e/JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh: line 373: 4529 Segmentation fault: 11 (core dumped) $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}" Return code: 1 Raw output file: /tmp/helix/working/A9B009AA/w/B9F709BA/uploads/Reports/JIT.HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.output.txt Raw output: BEGIN EXECUTION /tmp/helix/working/A9B009AA/p/corerun -p System.Reflection.Metadata.MetadataUpdater.IsSupported=false AdvSimd.Arm64_Part3_r.dll '' Supported ISAs: AdvSimd: True Aes: True ArmBase: True Crc32: True Dp: True Rdm: True Sha1: True Sha256: True Beginning test case MaxNumberPairwise.Vector128.Single at 7/24/2022 4:35:16 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScenario_Load Beginning scenario: RunReflectionScenario_UnsafeRead Beginning scenario: RunReflectionScenario_Load Beginning scenario: RunClsVarScenario Beginning scenario: RunClsVarScenario_Load Beginning scenario: RunLclVarScenario_UnsafeRead Beginning scenario: RunLclVarScenario_Load Beginning scenario: RunClassLclFldScenario Beginning scenario: RunClassLclFldScenario_Load Beginning scenario: RunClassFldScenario Beginning scenario: RunClassFldScenario_Load Beginning scenario: RunStructLclFldScenario Beginning scenario: RunStructLclFldScenario_Load Beginning scenario: RunStructFldScenario Beginning scenario: RunStructFldScenario_Load Ending test case at 7/24/2022 4:35:26 PM Beginning test case MaxNumberPairwiseScalar.Vector64.Single at 7/24/2022 4:35:26 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScenario_Load Beginning scenario: RunReflectionScenario_UnsafeRead Beginning scenario: RunReflectionScenario_Load Beginning scenario: RunClsVarScenario Beginning scenario: RunClsVarScenario_Load Beginning scenario: RunLclVarScenario_UnsafeRead Beginning scenario: RunLclVarScenario_Load Beginning scenario: RunClassLclFldScenario Beginning scenario: RunClassLclFldScenario_Load Beginning scenario: RunClassFldScenario Beginning scenario: RunClassFldScenario_Load Beginning scenario: RunStructLclFldScenario Beginning scenario: RunStructLclFldScenario_Load Beginning scenario: RunStructFldScenario Beginning scenario: RunStructFldScenario_Load Ending test case at 7/24/2022 4:35:28 PM Beginning test case MaxNumberPairwiseScalar.Vector128.Double at 7/24/2022 4:35:28 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScenario_Load Beginning scenario: RunReflectionScenario_UnsafeRead Beginning scenario: RunReflectionScenario_Load Beginning scenario: RunClsVarScenario Beginning scenario: RunClsVarScenario_Load Beginning scenario: RunLclVarScenario_UnsafeRead Beginning scenario: RunLclVarScenario_Load Beginning scenario: RunClassLclFldScenario Beginning scenario: RunClassLclFldScenario_Load Beginning scenario: RunClassFldScenario Beginning scenario: RunClassFldScenario_Load Beginning scenario: RunStructLclFldScenario Beginning scenario: RunStructLclFldScenario_Load Beginning scenario: RunStructFldScenario Beginning scenario: RunStructFldScenario_Load Ending test case at 7/24/2022 4:35:30 PM Beginning test case MaxPairwise.Vector128.Byte at 7/24/2022 4:35:30 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScen Stack trace at JIT_HardwareIntrinsics._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_sh() at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr) ```
1.0
Test failure JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh - Run: [runtime-coreclr gcstress-extra 20220724.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1900569&view=ms.vss-test-web.build-test-results-tab&runId=49457008&paneView=debug&resultId=108429) Failed test: ``` coreclr OSX arm64 Checked gcstress0xc_zapdisable_heapverify1 @ OSX.1200.ARM64.Open - JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh ``` **Error message:** ``` [createdump] Invalid process id: task_for_pid(4529) FAILED (os/kern) failure (5) [createdump] This failure may be because createdump or the application is not properly signed and entitled. [createdump] Failure took 0ms /private/tmp/helix/working/A9B009AA/w/B9F709BA/e/JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh: line 373: 4529 Segmentation fault: 11 (core dumped) $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}" Return code: 1 Raw output file: /tmp/helix/working/A9B009AA/w/B9F709BA/uploads/Reports/JIT.HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.output.txt Raw output: BEGIN EXECUTION /tmp/helix/working/A9B009AA/p/corerun -p System.Reflection.Metadata.MetadataUpdater.IsSupported=false AdvSimd.Arm64_Part3_r.dll '' Supported ISAs: AdvSimd: True Aes: True ArmBase: True Crc32: True Dp: True Rdm: True Sha1: True Sha256: True Beginning test case MaxNumberPairwise.Vector128.Single at 7/24/2022 4:35:16 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScenario_Load Beginning scenario: RunReflectionScenario_UnsafeRead Beginning scenario: RunReflectionScenario_Load Beginning scenario: RunClsVarScenario Beginning scenario: RunClsVarScenario_Load Beginning scenario: RunLclVarScenario_UnsafeRead Beginning scenario: RunLclVarScenario_Load Beginning scenario: RunClassLclFldScenario Beginning scenario: RunClassLclFldScenario_Load Beginning scenario: RunClassFldScenario Beginning scenario: RunClassFldScenario_Load Beginning scenario: RunStructLclFldScenario Beginning scenario: RunStructLclFldScenario_Load Beginning scenario: RunStructFldScenario Beginning scenario: RunStructFldScenario_Load Ending test case at 7/24/2022 4:35:26 PM Beginning test case MaxNumberPairwiseScalar.Vector64.Single at 7/24/2022 4:35:26 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScenario_Load Beginning scenario: RunReflectionScenario_UnsafeRead Beginning scenario: RunReflectionScenario_Load Beginning scenario: RunClsVarScenario Beginning scenario: RunClsVarScenario_Load Beginning scenario: RunLclVarScenario_UnsafeRead Beginning scenario: RunLclVarScenario_Load Beginning scenario: RunClassLclFldScenario Beginning scenario: RunClassLclFldScenario_Load Beginning scenario: RunClassFldScenario Beginning scenario: RunClassFldScenario_Load Beginning scenario: RunStructLclFldScenario Beginning scenario: RunStructLclFldScenario_Load Beginning scenario: RunStructFldScenario Beginning scenario: RunStructFldScenario_Load Ending test case at 7/24/2022 4:35:28 PM Beginning test case MaxNumberPairwiseScalar.Vector128.Double at 7/24/2022 4:35:28 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScenario_Load Beginning scenario: RunReflectionScenario_UnsafeRead Beginning scenario: RunReflectionScenario_Load Beginning scenario: RunClsVarScenario Beginning scenario: RunClsVarScenario_Load Beginning scenario: RunLclVarScenario_UnsafeRead Beginning scenario: RunLclVarScenario_Load Beginning scenario: RunClassLclFldScenario Beginning scenario: RunClassLclFldScenario_Load Beginning scenario: RunClassFldScenario Beginning scenario: RunClassFldScenario_Load Beginning scenario: RunStructLclFldScenario Beginning scenario: RunStructLclFldScenario_Load Beginning scenario: RunStructFldScenario Beginning scenario: RunStructFldScenario_Load Ending test case at 7/24/2022 4:35:30 PM Beginning test case MaxPairwise.Vector128.Byte at 7/24/2022 4:35:30 PM Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro Beginning scenario: RunBasicScenario_UnsafeRead Beginning scenario: RunBasicScen Stack trace at JIT_HardwareIntrinsics._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_sh() at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr) ```
non_process
test failure jit hardwareintrinsics arm advsimd advsimd r advsimd r sh run failed test coreclr osx checked zapdisable osx open jit hardwareintrinsics arm advsimd advsimd r advsimd r sh error message invalid process id task for pid failed os kern failure this failure may be because createdump or the application is not properly signed and entitled failure took private tmp helix working w e jit hardwareintrinsics arm advsimd advsimd r advsimd r sh line segmentation fault core dumped launcher exepath clrtestexecutionarguments return code raw output file tmp helix working w uploads reports jit hardwareintrinsics arm advsimd advsimd r advsimd r output txt raw output begin execution tmp helix working p corerun p system reflection metadata metadataupdater issupported false advsimd r dll supported isas advsimd true aes true armbase true true dp true rdm true true true beginning test case maxnumberpairwise single at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscenario load beginning scenario runreflectionscenario unsaferead beginning scenario runreflectionscenario load beginning scenario runclsvarscenario beginning scenario runclsvarscenario load beginning scenario runlclvarscenario unsaferead beginning scenario runlclvarscenario load beginning scenario runclasslclfldscenario beginning scenario runclasslclfldscenario load beginning scenario runclassfldscenario beginning scenario runclassfldscenario load beginning scenario runstructlclfldscenario beginning scenario runstructlclfldscenario load beginning scenario runstructfldscenario beginning scenario runstructfldscenario load ending test case at pm beginning test case maxnumberpairwisescalar single at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscenario load beginning scenario runreflectionscenario unsaferead beginning scenario runreflectionscenario load beginning scenario runclsvarscenario beginning scenario runclsvarscenario load beginning scenario runlclvarscenario unsaferead beginning scenario runlclvarscenario load beginning scenario runclasslclfldscenario beginning scenario runclasslclfldscenario load beginning scenario runclassfldscenario beginning scenario runclassfldscenario load beginning scenario runstructlclfldscenario beginning scenario runstructlclfldscenario load beginning scenario runstructfldscenario beginning scenario runstructfldscenario load ending test case at pm beginning test case maxnumberpairwisescalar double at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscenario load beginning scenario runreflectionscenario unsaferead beginning scenario runreflectionscenario load beginning scenario runclsvarscenario beginning scenario runclsvarscenario load beginning scenario runlclvarscenario unsaferead beginning scenario runlclvarscenario load beginning scenario runclasslclfldscenario beginning scenario runclasslclfldscenario load beginning scenario runclassfldscenario beginning scenario runclassfldscenario load beginning scenario runstructlclfldscenario beginning scenario runstructlclfldscenario load beginning scenario runstructfldscenario beginning scenario runstructfldscenario load ending test case at pm beginning test case maxpairwise byte at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscen stack trace at jit hardwareintrinsics arm advsimd advsimd r advsimd r arm advsimd advsimd r advsimd r sh at system runtimemethodhandle invokemethod object target void arguments signature sig boolean isconstructor at system reflection methodinvoker invoke object obj intptr args bindingflags invokeattr
0
11,757
14,591,552,135
IssuesEvent
2020-12-19 13:34:16
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
symfony/process returns empty outputs on IIS
Bug Process Status: Needs Review
| Q | A | ---------------- | ----- | Bug report? | yes | Feature request? | no | BC Break report? | no | RFC? | no | Symfony version | 3.3.11 Hello, I'm using PHP 7.1 on IIS 7.5 and symfony/process always returns an empty output. Here is a test case: ```php <?php use Symfony\Component\Process\Process; require_once __DIR__.'/vendor/autoload.php'; //This returns an output var_dump(shell_exec('dir')); $process = new Process('dir'); $process->mustRun(); //These return empty strings var_dump( $process->getOutput(), $process->getErrorOutput() ); ``` What's strange is that it works when calling PHP (same binary) from the commandline. And of course the same code works fine on my Linux/Apache server. I tested various other commands (`cd`, `Python.exe`) and I always get the same empty result. `mustRun()` does throw an exception if the command does not exist: ``` PHP Fatal error: Uncaught Symfony\Component\Process\Exception\ProcessFailedException: The command "foobar" failed. Exit Code: 1(General error) Working directory: C:\inetpub\wwwroot Output: ================ Error Output: ================ in C:\inetpub\wwwroot\vendor\symfony\process\Process.php:241 Stack trace: #0 C:\inetpub\wwwroot\test.php(11): Symfony\Component\Process\Process->mustRun() #1 {main} thrown in C:\inetpub\wwwroot\vendor\symfony\process\Process.php on line 241 ``` Edit: I also tried configuring IIS to use another PHP binary (installed with Chocolatey) and I get the same issue, so I guess it is linked to IIS itself.
1.0
symfony/process returns empty outputs on IIS - | Q | A | ---------------- | ----- | Bug report? | yes | Feature request? | no | BC Break report? | no | RFC? | no | Symfony version | 3.3.11 Hello, I'm using PHP 7.1 on IIS 7.5 and symfony/process always returns an empty output. Here is a test case: ```php <?php use Symfony\Component\Process\Process; require_once __DIR__.'/vendor/autoload.php'; //This returns an output var_dump(shell_exec('dir')); $process = new Process('dir'); $process->mustRun(); //These return empty strings var_dump( $process->getOutput(), $process->getErrorOutput() ); ``` What's strange is that it works when calling PHP (same binary) from the commandline. And of course the same code works fine on my Linux/Apache server. I tested various other commands (`cd`, `Python.exe`) and I always get the same empty result. `mustRun()` does throw an exception if the command does not exist: ``` PHP Fatal error: Uncaught Symfony\Component\Process\Exception\ProcessFailedException: The command "foobar" failed. Exit Code: 1(General error) Working directory: C:\inetpub\wwwroot Output: ================ Error Output: ================ in C:\inetpub\wwwroot\vendor\symfony\process\Process.php:241 Stack trace: #0 C:\inetpub\wwwroot\test.php(11): Symfony\Component\Process\Process->mustRun() #1 {main} thrown in C:\inetpub\wwwroot\vendor\symfony\process\Process.php on line 241 ``` Edit: I also tried configuring IIS to use another PHP binary (installed with Chocolatey) and I get the same issue, so I guess it is linked to IIS itself.
process
symfony process returns empty outputs on iis q a bug report yes feature request no bc break report no rfc no symfony version hello i m using php on iis and symfony process always returns an empty output here is a test case php php use symfony component process process require once dir vendor autoload php this returns an output var dump shell exec dir process new process dir process mustrun these return empty strings var dump process getoutput process geterroroutput what s strange is that it works when calling php same binary from the commandline and of course the same code works fine on my linux apache server i tested various other commands cd python exe and i always get the same empty result mustrun does throw an exception if the command does not exist php fatal error uncaught symfony component process exception processfailedexception the command foobar failed exit code general error working directory c inetpub wwwroot output error output in c inetpub wwwroot vendor symfony process process php stack trace c inetpub wwwroot test php symfony component process process mustrun main thrown in c inetpub wwwroot vendor symfony process process php on line edit i also tried configuring iis to use another php binary installed with chocolatey and i get the same issue so i guess it is linked to iis itself
1
92,620
8,373,219,296
IssuesEvent
2018-10-05 09:39:34
andrewwood2/acebook-gazelle
https://api.github.com/repos/andrewwood2/acebook-gazelle
closed
B. Set up front end feature testing framework
in progress test
Set up feature testing for React front end. either Mocha, Zombie, etc.
1.0
B. Set up front end feature testing framework - Set up feature testing for React front end. either Mocha, Zombie, etc.
non_process
b set up front end feature testing framework set up feature testing for react front end either mocha zombie etc
0
41,197
10,331,326,860
IssuesEvent
2019-09-02 17:33:01
davidjamesca/ctypesgen
https://api.github.com/repos/davidjamesca/ctypesgen
closed
Tests failing to find libc.so.6 and libm.so.6 on 64bit Ubuntu (SVN r147)
Priority-Medium Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Checkout the code 2. Go into the test directory 3. Run "./testsuite.py" What is the expected output? What do you see instead? I expect to see all tests pass instead I see 11 errors all like this: ====================================================================== ERROR: test_bad_args_string_not_number (__main__.MathTest) Based on math_functions.py ---------------------------------------------------------------------- Traceback (most recent call last): File "./testsuite.py", line 252, in setUp self.module, output = ctypesgentest.test(header_str, libraries=libraries, all_headers=True) File "/home/jlisee/projects/ctypesgen-read-only/test/ctypesgentest.py", line 52, in test module = __import__("temp") File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 598, in <module> _libs["libm.so.6"] = load_library("libm.so.6") File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 367, in load_library raise ImportError("%s not found." % libname) ImportError: libm.so.6 not found What version of the product are you using? On what operating system? Ubuntu 12.04 64bit, SVN r147. Please provide any additional information below. I have attached a patch to fix the issue. There are still 3 tests failing with "AttributeError: type object 'c_uint' has no attribute '_fields_'" in the generated "temp.py" file. ``` Original issue reported on code.google.com by `jli...@gmail.com` on 28 Feb 2013 at 3:02 Attachments: - [Ubuntu_64bit_fix.patch](https://storage.googleapis.com/google-code-attachments/ctypesgen/issue-39/comment-0/Ubuntu_64bit_fix.patch)
1.0
Tests failing to find libc.so.6 and libm.so.6 on 64bit Ubuntu (SVN r147) - ``` What steps will reproduce the problem? 1. Checkout the code 2. Go into the test directory 3. Run "./testsuite.py" What is the expected output? What do you see instead? I expect to see all tests pass instead I see 11 errors all like this: ====================================================================== ERROR: test_bad_args_string_not_number (__main__.MathTest) Based on math_functions.py ---------------------------------------------------------------------- Traceback (most recent call last): File "./testsuite.py", line 252, in setUp self.module, output = ctypesgentest.test(header_str, libraries=libraries, all_headers=True) File "/home/jlisee/projects/ctypesgen-read-only/test/ctypesgentest.py", line 52, in test module = __import__("temp") File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 598, in <module> _libs["libm.so.6"] = load_library("libm.so.6") File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 367, in load_library raise ImportError("%s not found." % libname) ImportError: libm.so.6 not found What version of the product are you using? On what operating system? Ubuntu 12.04 64bit, SVN r147. Please provide any additional information below. I have attached a patch to fix the issue. There are still 3 tests failing with "AttributeError: type object 'c_uint' has no attribute '_fields_'" in the generated "temp.py" file. ``` Original issue reported on code.google.com by `jli...@gmail.com` on 28 Feb 2013 at 3:02 Attachments: - [Ubuntu_64bit_fix.patch](https://storage.googleapis.com/google-code-attachments/ctypesgen/issue-39/comment-0/Ubuntu_64bit_fix.patch)
non_process
tests failing to find libc so and libm so on ubuntu svn what steps will reproduce the problem checkout the code go into the test directory run testsuite py what is the expected output what do you see instead i expect to see all tests pass instead i see errors all like this error test bad args string not number main mathtest based on math functions py traceback most recent call last file testsuite py line in setup self module output ctypesgentest test header str libraries libraries all headers true file home jlisee projects ctypesgen read only test ctypesgentest py line in test module import temp file home jlisee projects ctypesgen read only test temp py line in libs load library libm so file home jlisee projects ctypesgen read only test temp py line in load library raise importerror s not found libname importerror libm so not found what version of the product are you using on what operating system ubuntu svn please provide any additional information below i have attached a patch to fix the issue there are still tests failing with attributeerror type object c uint has no attribute fields in the generated temp py file original issue reported on code google com by jli gmail com on feb at attachments
0
757,685
26,524,390,024
IssuesEvent
2023-01-19 07:18:54
pystardust/ani-cli
https://api.github.com/repos/pystardust/ani-cli
closed
Episodes not released yet!
type: bug priority 2: medium
Version: 3.4.0 OS: Debian 11 Shell: Bash 5.1.4 Anime: JoJo I get an error "Episodes not released yet!" no matter the anime I choose 1. Run `ani-cli JoJo 2. Choose 1 3. Episodes not released yet!
1.0
Episodes not released yet! - Version: 3.4.0 OS: Debian 11 Shell: Bash 5.1.4 Anime: JoJo I get an error "Episodes not released yet!" no matter the anime I choose 1. Run `ani-cli JoJo 2. Choose 1 3. Episodes not released yet!
non_process
episodes not released yet version os debian shell bash anime jojo i get an error episodes not released yet no matter the anime i choose run ani cli jojo choose episodes not released yet
0
276,637
20,993,384,598
IssuesEvent
2022-03-29 11:25:21
Sitecore/developer-portal
https://api.github.com/repos/Sitecore/developer-portal
closed
Create DevOps guide for Managed Cloud Containers
documentation
Need an article added to the Getting Started section (`/learn/getting-started`) area that can pull together all the documentation and steps a developer may need to get started with container GitOps for Managed Cloud Containers. We also don't have a Managed Cloud product area on the site yet, so that will need to be added as well so that this guide can be found.
1.0
Create DevOps guide for Managed Cloud Containers - Need an article added to the Getting Started section (`/learn/getting-started`) area that can pull together all the documentation and steps a developer may need to get started with container GitOps for Managed Cloud Containers. We also don't have a Managed Cloud product area on the site yet, so that will need to be added as well so that this guide can be found.
non_process
create devops guide for managed cloud containers need an article added to the getting started section learn getting started area that can pull together all the documentation and steps a developer may need to get started with container gitops for managed cloud containers we also don t have a managed cloud product area on the site yet so that will need to be added as well so that this guide can be found
0
10,920
13,697,018,917
IssuesEvent
2020-10-01 01:46:37
opendistro-for-elasticsearch/opendistro-build
https://api.github.com/repos/opendistro-for-elasticsearch/opendistro-build
closed
Release ODFE 1.10.1 based on ES 7.9.1
in process infra new release
ODFE 1.10.1 based on ES 7.9.1 **(Note: ODFE 1.10.0 is skipped right now as we prefer to release ODFE 1.10.1 for ES 7.9.1. This is to avoid memory leak in Lucene 8.6.0 and 8.6.1 (ES 7.9.1 has Lucene 8.6.2) https://github.com/elastic/elasticsearch/issues/61512)** Release Engineering / Build Repo Key Changes: * KNNLib will now use wildcard to resolve hardcoded version issues ([#359](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/359)) * Docker allows elasticsearch user to access logs under supervisord folder ([#271](https://github.com/opendistro-for-elasticsearch/performance-analyzer-rca/pull/271), [#146](https://github.com/opendistro-for-elasticsearch/performance-analyzer/pull/146), [#320](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/320)) * Implement Version Cuts for consistent distribution release builds ([#357](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/357)) * Add descriptions for several scripts with usage documentations ([#334](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/334)) * Update opendistro-build github repo issues link ([#382](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/382)) * Disable optimizations for KNNLib compilation in docker image creation ([#384](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/384)) * HELM allows customizing docker registry, thanks @tareqhs ([#358](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/358)) * HELM Kibana ingress path fix, thanks @Hokwang ([#340](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/340)) * Helm master nodes allows extraVolumeMounts when securityconfig disabled, thanks @aplhk ([#366](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/366)) * HELM Readme Update, thanks @dmpe ([#380](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/380) [#385](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/385)) * Kibana has new cookie settings for security kibana plugin 2.0 framework ([#397](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/397)) ------ ### ODFE 1.10.1 Released as of 2020/09/30 * Downloads [Here](https://opendistro.github.io/for-elasticsearch/downloads.html) * [1.10.1 Release Notes](https://github.com/opendistro-for-elasticsearch/opendistro-build/blob/master/release-notes/opendistro-for-elasticsearch-release-notes-1.10.1.md) * [1.10.1 Blog Post](https://opendistro.github.io/for-elasticsearch/blog/odfe-updates/2020/09/Open-Distro-for-Elasticsearch-1.10.1-is-released/) ------ ODFE 1.10.0 post for backup purposes: https://github.com/opendistro-for-elasticsearch/opendistro-build/issues/350
1.0
Release ODFE 1.10.1 based on ES 7.9.1 - ODFE 1.10.1 based on ES 7.9.1 **(Note: ODFE 1.10.0 is skipped right now as we prefer to release ODFE 1.10.1 for ES 7.9.1. This is to avoid memory leak in Lucene 8.6.0 and 8.6.1 (ES 7.9.1 has Lucene 8.6.2) https://github.com/elastic/elasticsearch/issues/61512)** Release Engineering / Build Repo Key Changes: * KNNLib will now use wildcard to resolve hardcoded version issues ([#359](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/359)) * Docker allows elasticsearch user to access logs under supervisord folder ([#271](https://github.com/opendistro-for-elasticsearch/performance-analyzer-rca/pull/271), [#146](https://github.com/opendistro-for-elasticsearch/performance-analyzer/pull/146), [#320](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/320)) * Implement Version Cuts for consistent distribution release builds ([#357](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/357)) * Add descriptions for several scripts with usage documentations ([#334](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/334)) * Update opendistro-build github repo issues link ([#382](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/382)) * Disable optimizations for KNNLib compilation in docker image creation ([#384](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/384)) * HELM allows customizing docker registry, thanks @tareqhs ([#358](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/358)) * HELM Kibana ingress path fix, thanks @Hokwang ([#340](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/340)) * Helm master nodes allows extraVolumeMounts when securityconfig disabled, thanks @aplhk ([#366](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/366)) * HELM Readme Update, thanks @dmpe ([#380](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/380) [#385](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/385)) * Kibana has new cookie settings for security kibana plugin 2.0 framework ([#397](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/397)) ------ ### ODFE 1.10.1 Released as of 2020/09/30 * Downloads [Here](https://opendistro.github.io/for-elasticsearch/downloads.html) * [1.10.1 Release Notes](https://github.com/opendistro-for-elasticsearch/opendistro-build/blob/master/release-notes/opendistro-for-elasticsearch-release-notes-1.10.1.md) * [1.10.1 Blog Post](https://opendistro.github.io/for-elasticsearch/blog/odfe-updates/2020/09/Open-Distro-for-Elasticsearch-1.10.1-is-released/) ------ ODFE 1.10.0 post for backup purposes: https://github.com/opendistro-for-elasticsearch/opendistro-build/issues/350
process
release odfe based on es odfe based on es note odfe is skipped right now as we prefer to release odfe for es this is to avoid memory leak in lucene and es has lucene release engineering build repo key changes knnlib will now use wildcard to resolve hardcoded version issues docker allows elasticsearch user to access logs under supervisord folder implement version cuts for consistent distribution release builds add descriptions for several scripts with usage documentations update opendistro build github repo issues link disable optimizations for knnlib compilation in docker image creation helm allows customizing docker registry thanks tareqhs helm kibana ingress path fix thanks hokwang helm master nodes allows extravolumemounts when securityconfig disabled thanks aplhk helm readme update thanks dmpe kibana has new cookie settings for security kibana plugin framework odfe released as of downloads odfe post for backup purposes
1
69,037
7,122,685,713
IssuesEvent
2018-01-19 12:49:36
nodejs/node
https://api.github.com/repos/nodejs/node
opened
Flaky http2-settings-flood
CI / flaky test freebsd
Multiple timeouts on this test: https://ci.nodejs.org/job/node-test-commit-freebsd/14756/nodes=freebsd10-64/console https://ci.nodejs.org/job/node-test-commit-freebsd/14746/nodes=freebsd10-64/console ``` not ok 2035 sequential/test-http2-settings-flood --- duration_ms: 120.105 severity: fail stack: |- timeout ```
1.0
Flaky http2-settings-flood - Multiple timeouts on this test: https://ci.nodejs.org/job/node-test-commit-freebsd/14756/nodes=freebsd10-64/console https://ci.nodejs.org/job/node-test-commit-freebsd/14746/nodes=freebsd10-64/console ``` not ok 2035 sequential/test-http2-settings-flood --- duration_ms: 120.105 severity: fail stack: |- timeout ```
non_process
flaky settings flood multiple timeouts on this test not ok sequential test settings flood duration ms severity fail stack timeout
0
618,568
19,474,854,104
IssuesEvent
2021-12-24 10:07:40
MartinXPN/profound.academy
https://api.github.com/repos/MartinXPN/profound.academy
closed
Allow visitors to view exercises
enhancement priority/top
Allow visitors to view exercises but not submit them until they sign up. In case they submit anything add the course to curriculum
1.0
Allow visitors to view exercises - Allow visitors to view exercises but not submit them until they sign up. In case they submit anything add the course to curriculum
non_process
allow visitors to view exercises allow visitors to view exercises but not submit them until they sign up in case they submit anything add the course to curriculum
0
19,563
25,885,019,166
IssuesEvent
2022-12-14 13:59:51
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/spanmetrics] getting key from cache error
bug processor/spanmetrics
### What happened? ## Description I get an error in my app when it sent span to a otel collector which configed a spanmetrics processor with `defaultDimensionsCacheSize`: ``` 2022/10/27 02:37:46 rpc error: code = Unknown desc = value not found in metricKeyToDimensions cache by key "amamba\x00amamba.io.api.pipeline.v1alpha1.Pipelines/ReplayPipelineRun\x00SPAN_KIND_SERVER\x00STATUS_CODE_OK\x00cd7b102e-fbc5-4556-a0fd-718298df3de9\x00amamba-system\x00demo-dev-worker-03\x00amamba-apiserver-66c9486b55-5l4n8" ``` I trick the error to: https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/92ad54f707a967386b764b6fdaea5a2ba4377319/processor/spanmetricsprocessor/processor.go#L335-L344 ### Collector version v0.61.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration _No response_ ### Log output _No response_ ### Additional context resolve by #15687
1.0
[processor/spanmetrics] getting key from cache error - ### What happened? ## Description I get an error in my app when it sent span to a otel collector which configed a spanmetrics processor with `defaultDimensionsCacheSize`: ``` 2022/10/27 02:37:46 rpc error: code = Unknown desc = value not found in metricKeyToDimensions cache by key "amamba\x00amamba.io.api.pipeline.v1alpha1.Pipelines/ReplayPipelineRun\x00SPAN_KIND_SERVER\x00STATUS_CODE_OK\x00cd7b102e-fbc5-4556-a0fd-718298df3de9\x00amamba-system\x00demo-dev-worker-03\x00amamba-apiserver-66c9486b55-5l4n8" ``` I trick the error to: https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/92ad54f707a967386b764b6fdaea5a2ba4377319/processor/spanmetricsprocessor/processor.go#L335-L344 ### Collector version v0.61.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration _No response_ ### Log output _No response_ ### Additional context resolve by #15687
process
getting key from cache error what happened description i get an error in my app when it sent span to a otel collector which configed a spanmetrics processor with defaultdimensionscachesize rpc error code unknown desc value not found in metrickeytodimensions cache by key amamba io api pipeline pipelines replaypipelinerun kind server code ok system dev worker apiserver i trick the error to collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration no response log output no response additional context resolve by
1
325,176
9,920,360,218
IssuesEvent
2019-06-30 08:30:45
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
closed
Synthesis failed for DLP
:rotating_light: autosynth failure priority: p1 triage me type: bug
Hello! Autosynth couldn't regenerate DLP. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-dlp' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/dlp/synth.metadata', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. synthtool > Failed executing git clean -fdx: warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/reviews.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/edits.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/orders.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/purchases.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/inappproducts.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/internalappsharingartifacts.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/connection.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundles_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_file.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundle.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/image.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk_binary.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listing.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/page_info.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/testers.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/season.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_details.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review_reply_result.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/tracks_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchase.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchase.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchases_acknowledge_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_cancel_survey_result.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/timestamp.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_files_upload_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_upload_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_price_change.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_deferral_info.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_acknowledge_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchase.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_delete_all_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/price.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/prorate.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk_uses_permission.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/developer_comment.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchases_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_file.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/token_pagination.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/localized_text.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/country_targeting.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/month_day.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_files_upload_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track_release.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/device_metadata.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product_listing.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_edit.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/internal_app_sharing_artifact.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/inappproducts_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/user_comment.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listings_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/comment.ex: Permission denied Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 32, in <module> shell.run(["git", "clean", "-fdx"], cwd=repository / "clients") File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1. synthtool > Wrote metadata to clients/dlp/synth.metadata. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/6d24954b-bcff-4c08-9c3d-47b4d58021a3).
1.0
Synthesis failed for DLP - Hello! Autosynth couldn't regenerate DLP. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-dlp' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/dlp/synth.metadata', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. synthtool > Failed executing git clean -fdx: warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/reviews.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/edits.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/orders.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/purchases.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/inappproducts.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/internalappsharingartifacts.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/connection.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundles_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_file.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundle.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/image.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk_binary.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listing.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/page_info.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/testers.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/season.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_details.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review_reply_result.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/tracks_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchase.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchase.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchases_acknowledge_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_cancel_survey_result.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/timestamp.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_files_upload_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_upload_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_price_change.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_deferral_info.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_acknowledge_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchase.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_delete_all_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/price.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/prorate.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk_uses_permission.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/developer_comment.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchases_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_file.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/token_pagination.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/localized_text.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/country_targeting.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/month_day.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_files_upload_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_request.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track_release.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/device_metadata.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product_listing.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_edit.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/internal_app_sharing_artifact.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/inappproducts_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/user_comment.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listings_list_response.ex: Permission denied warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/comment.ex: Permission denied Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 32, in <module> shell.run(["git", "clean", "-fdx"], cwd=repository / "clients") File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1. synthtool > Wrote metadata to clients/dlp/synth.metadata. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/6d24954b-bcff-4c08-9c3d-47b4d58021a3).
non_process
synthesis failed for dlp hello autosynth couldn t regenerate dlp broken heart here s the output from running synth py cloning into working repo switched to branch autosynth dlp running synthtool synthtool executing tmpfs src git autosynth working repo synth py synthtool cloning synthtool failed executing git clean fdx warning failed to remove android publisher lib google api android publisher api reviews ex permission denied warning failed to remove android publisher lib google api android publisher api edits ex permission denied warning failed to remove android publisher lib google api android publisher api orders ex permission denied warning failed to remove android publisher lib google api android publisher api purchases ex permission denied warning failed to remove android publisher lib google api android publisher api inappproducts ex permission denied warning failed to remove android publisher lib google api android publisher api internalappsharingartifacts ex permission denied warning failed to remove android publisher lib google api android publisher connection ex permission denied warning failed to remove android publisher lib google api android publisher model bundles list response ex permission denied warning failed to remove android publisher lib google api android publisher model deobfuscation file ex permission denied warning failed to remove android publisher lib google api android publisher model bundle ex permission denied warning failed to remove android publisher lib google api android publisher model image ex permission denied warning failed to remove android publisher lib google api android publisher model apk binary ex permission denied warning failed to remove android publisher lib google api android publisher model listing ex permission denied warning failed to remove android publisher lib google api android publisher model page info ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchases defer request ex permission denied warning failed to remove android publisher lib google api android publisher model testers ex permission denied warning failed to remove android publisher lib google api android publisher model track ex permission denied warning failed to remove android publisher lib google api android publisher model in app product ex permission denied warning failed to remove android publisher lib google api android publisher model season ex permission denied warning failed to remove android publisher lib google api android publisher model app details ex permission denied warning failed to remove android publisher lib google api android publisher model apks list response ex permission denied warning failed to remove android publisher lib google api android publisher model review reply result ex permission denied warning failed to remove android publisher lib google api android publisher model tracks list response ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchases defer response ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchase ex permission denied warning failed to remove android publisher lib google api android publisher model product purchase ex permission denied warning failed to remove android publisher lib google api android publisher model product purchases acknowledge request ex permission denied warning failed to remove android publisher lib google api android publisher model subscription cancel survey result ex permission denied warning failed to remove android publisher lib google api android publisher model timestamp ex permission denied warning failed to remove android publisher lib google api android publisher model deobfuscation files upload response ex permission denied warning failed to remove android publisher lib google api android publisher model images upload response ex permission denied warning failed to remove android publisher lib google api android publisher model subscription price change ex permission denied warning failed to remove android publisher lib google api android publisher model subscription deferral info ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchases acknowledge request ex permission denied warning failed to remove android publisher lib google api android publisher model reviews list response ex permission denied warning failed to remove android publisher lib google api android publisher model voided purchase ex permission denied warning failed to remove android publisher lib google api android publisher model images delete all response ex permission denied warning failed to remove android publisher lib google api android publisher model price ex permission denied warning failed to remove android publisher lib google api android publisher model reviews reply request ex permission denied warning failed to remove android publisher lib google api android publisher model prorate ex permission denied warning failed to remove android publisher lib google api android publisher model externally hosted apk uses permission ex permission denied warning failed to remove android publisher lib google api android publisher model developer comment ex permission denied warning failed to remove android publisher lib google api android publisher model images list response ex permission denied warning failed to remove android publisher lib google api android publisher model voided purchases list response ex permission denied warning failed to remove android publisher lib google api android publisher model review ex permission denied warning failed to remove android publisher lib google api android publisher model expansion file ex permission denied warning failed to remove android publisher lib google api android publisher model token pagination ex permission denied warning failed to remove android publisher lib google api android publisher model apk ex permission denied warning failed to remove android publisher lib google api android publisher model localized text ex permission denied warning failed to remove android publisher lib google api android publisher model country targeting ex permission denied warning failed to remove android publisher lib google api android publisher model month day ex permission denied warning failed to remove android publisher lib google api android publisher model apks add externally hosted response ex permission denied warning failed to remove android publisher lib google api android publisher model reviews reply response ex permission denied warning failed to remove android publisher lib google api android publisher model expansion files upload response ex permission denied warning failed to remove android publisher lib google api android publisher model externally hosted apk ex permission denied warning failed to remove android publisher lib google api android publisher model apks add externally hosted request ex permission denied warning failed to remove android publisher lib google api android publisher model track release ex permission denied warning failed to remove android publisher lib google api android publisher model device metadata ex permission denied warning failed to remove android publisher lib google api android publisher model in app product listing ex permission denied warning failed to remove android publisher lib google api android publisher model app edit ex permission denied warning failed to remove android publisher lib google api android publisher model internal app sharing artifact ex permission denied warning failed to remove android publisher lib google api android publisher model inappproducts list response ex permission denied warning failed to remove android publisher lib google api android publisher model user comment ex permission denied warning failed to remove android publisher lib google api android publisher model listings list response ex permission denied warning failed to remove android publisher lib google api android publisher model comment ex permission denied traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth env lib site packages synthtool main py line in main file tmpfs src git autosynth env lib site packages click core py line in call return self main args kwargs file tmpfs src git autosynth env lib site packages click core py line in main rv self invoke ctx file tmpfs src git autosynth env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src git autosynth env lib site packages click core py line in invoke return callback args kwargs file tmpfs src git autosynth env lib site packages synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src git autosynth working repo synth py line in shell run cwd repository clients file tmpfs src git autosynth env lib site packages synthtool shell py line in run raise exc file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthtool wrote metadata to clients dlp synth metadata synthesis failed google internal developers can see the full log
0
18,179
24,231,196,227
IssuesEvent
2022-09-26 18:26:49
google/android-fhir
https://api.github.com/repos/google/android-fhir
closed
Support Android Min API Level 24
process
We have not been testing API level 21 consistently in our testing matrix: https://github.com/google/android-fhir/blob/master/.github/workflows/device-tests.yml#L40 We should consider raising the min api level to 24 (released aug 2016) as we already do not guarantee support for anything lower than 24 due to our lack of tests.
1.0
Support Android Min API Level 24 - We have not been testing API level 21 consistently in our testing matrix: https://github.com/google/android-fhir/blob/master/.github/workflows/device-tests.yml#L40 We should consider raising the min api level to 24 (released aug 2016) as we already do not guarantee support for anything lower than 24 due to our lack of tests.
process
support android min api level we have not been testing api level consistently in our testing matrix we should consider raising the min api level to released aug as we already do not guarantee support for anything lower than due to our lack of tests
1
94,648
27,253,019,431
IssuesEvent
2023-02-22 09:32:20
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Godot 4 compiled with MinGW is crashing with a white screen [GCC 11.x bug fixed in 11.3]
bug platform:windows topic:buildsystem topic:thirdparty crash
### Godot version 4.0 custom build (bdcc8741e) ### System information Windows 10 21H1 ### Issue description The editor is crashing with a white screen on startup, with or without a project. I have been tried some combinations of compilers and options: MSVC with debug -> OK MSVC with release_debug -> OK MinGW with debug -> OK MinGW with release_debug -> Crash This issue has been introduced with the 99fe462452be44efa618e83ad9bbecd722ae6ecd commit. <details> <summary>Output of drmingw</summary> ``` godot.windows.opt.tools.64.exe caused an Access Violation at location 00007FF73A4C0000 DEP violation at location 00007FF73A4C0000. AddrPC Params 00007FF73A4C0000 00000215616FA200 0000003CE61FF5A8 0000003CE61FF5B0 00007FF77C559842 0000000000400000 00007FF77CA33F03 0000000000400000 godot.windows.opt.tools.64.exe!RenderingServerDefault::RenderingServerDefault [E:/path/core/os/thread.h @ 90] 88: _FORCE_INLINE_ ID get_id() const { return id; } 89: // get the ID of the caller thread > 90: _FORCE_INLINE_ static ID get_caller_id() { return caller_id; } 91: // get the ID of the main thread 92: _FORCE_INLINE_ static ID get_main_id() { return main_thread_id; } 00007FF77A4DE90B 0000021561579CC0 0000021560C1F690 00000215000000FF godot.windows.opt.tools.64.exe!Main::setup2 [E:/path/main/main.cpp @ 1590] 1588: /* Initialize Rendering Server */ 1589: > 1590: rendering_server = memnew(RenderingServerDefault(OS::get_singleton()->get_render_thread_mode() == OS::RENDER_SEPARATE_THREAD)); 1591: 1592: rendering_server->init(); 00007FF77A4E479B 0000021560C20C30 0000000000000001 000002155F450860 godot.windows.opt.tools.64.exe!Main::setup [E:/path/main/main.cpp @ 1388] 1386: 1387: if (p_second_phase) { > 1388: return setup2(); 1389: } 1390: 00007FF77A4C18A7 000002155F153296 0000003CE61FFD4C 0000000000000001 godot.windows.opt.tools.64.exe!widechar_main [E:/path/platform/windows/godot_windows.cpp @ 151] 149: TEST_MAIN_PARAM_OVERRIDE(argc, argv_utf8) 150: > 151: Error err = Main::setup(argv_utf8[0], argc - 1, &argv_utf8[1]); 152: 153: if (err != OK) { 00007FF77A4C19A0 0000000000000000 0000000000000047 00007FF7811164D8 godot.windows.opt.tools.64.exe!_main [E:/path/platform/windows/godot_windows.cpp @ 185] 183: } 184: > 185: result = widechar_main(argc, wc_argv); 186: 187: LocalFree(wc_argv); 00007FF77A4C13B1 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!__tmainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 321] 00007FF77A4C14E6 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!mainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 202] 00007FF8F8997034 0000000000000000 0000000000000000 0000000000000000 KERNEL32.DLL!BaseThreadInitThunk 00007FF8FA1C2651 0000000000000000 0000000000000000 0000000000000000 ntdll.dll!RtlUserThreadStart ``` </details> <details> <summary>Log of the crashing MinGW release_debug build</summary> ``` Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org Using "ICU / HarfBuzz / Graphite" text server... Vulkan API 1.2.162 ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157) WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files. Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154) Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080 - Vulkan multiview supported: max view count: 32 max instances: 134217727 - Vulkan subgroup: size: 32 stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV quad operations in all stages Using present mode: VK_PRESENT_MODE_FIFO_KHR Using "winink" pen tablet driver... ``` </details> <details> <summary>Log of the working MinGW debug build</summary> ``` Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org Using "ICU / HarfBuzz / Graphite" text server... Vulkan API 1.2.162 ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157) WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files. Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154) Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080 - Vulkan multiview supported: max view count: 32 max instances: 134217727 - Vulkan subgroup: size: 32 stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV quad operations in all stages Using present mode: VK_PRESENT_MODE_FIFO_KHR Using "winink" pen tablet driver... Shader 'VoxelGiSdfShaderRD' SHA256: b6972a55781e21cfbac0dfc035381ee1f65e972c1f7e0bcc6004bfb74cc9dc9e Shader 'ParticlesShaderRD' SHA256: cbb6a5ce1933875a8e2df4f0cde4684286dedb6737cfbcfdeba745c978a2cc45 Shader 'ParticlesCopyShaderRD' SHA256: 6bd2ea6fca4c3a51f1e9bcd5068315e63820ec68fd8596e20480174c943bb9df Shader 'CanvasSdfShaderRD' SHA256: 692328d90ee7d404503d304e2464af3f3ac466ebd287762fd8883248f3e2a046 Shader 'SkeletonShaderRD' SHA256: 8f03a2ebc895d833120c3b2341c16e971fb1404dd3ae9e8030bc4bbf63e67fd5 Shader 'CanvasShaderRD' SHA256: a30a00a53b17f4e413dfb1fe109868063f2d67900aaeec66b0f50964108e216b Shader 'CanvasOcclusionShaderRD' SHA256: fd08d5c92b6537852fb4b8e5fba27f28518faad41a5c21ac9bd51bf49807824c Shader 'ClusterRenderShaderRD' SHA256: a2924afb04064a1dc42ed9d63721e7d5dd0466f245037ff85b5bef8bd251f515 Shader 'ClusterStoreShaderRD' SHA256: 64c437e6bebba9faa4766b524281310d934fd2a28f26e9a8a7834ca07f09d2dc Shader 'ClusterDebugShaderRD' SHA256: 0f66ac91d81e6f440ffa804d473579cb5d7ac135b375dce6d120186a65b54204 Shader 'SkyShaderRD' SHA256: dd0dbf614e400cc2f58360d292f85f19237fdd9746b9a46c6fd7c50fe8a2ae43 Shader 'VoxelGiShaderRD' SHA256: 51aa13996be8fb6071160d3ea73d752edcd67ef81e1aec1ec3623a13eee94d0a Shader 'VoxelGiDebugShaderRD' SHA256: 7fa6be49a459f8891a6e756165f2e9ca7f3ba261f53edf36c2041dc2acbf2a96 Shader 'SdfgiPreprocessShaderRD' SHA256: 056a9e6eaae9cc98093ceec6142bcb4582be0f3443e18402c1e25c293899b104 Shader 'SdfgiDirectLightShaderRD' SHA256: 0707e534eb11aeb7566a94913fe168d96b66896b440650f768f92ec6ed5cc842 Shader 'SdfgiIntegrateShaderRD' SHA256: deacba90b0c075f22f4000295f6a0b846957d2602b5d54b631c089792288e7df Shader 'GiShaderRD' SHA256: c1d65ee198e82b396759752f26b64ae98d7bb8691aed1bc07924617cf25872c5 Shader 'SdfgiDebugShaderRD' SHA256: 5e9f1e4d567be5c04dacec2ddbddb98638086a1fe7dc1d0260f3a58e93856a0e Shader 'SdfgiDebugProbesShaderRD' SHA256: a6a825191c0869482f513376305eb79fa16189e65068bf76a484a0e1eb602c1b Shader 'VolumetricFogShaderRD' SHA256: bc8fa35c7d92fd0353d5458af7d19438bf0c46f8c873b1b5e45f328f988f2d34 Shader 'SceneForwardClusteredShaderRD' SHA256: 300732714cbaf090328e341566d4d056b1bf83a2786d31de56778f6c929d1987 Shader 'CopyShaderRD' SHA256: 4ea51535b5cdcba7ece20ebd074657d05cbde47d1d0e87d3c84a2a1c28aa976f Shader 'CopyToFbShaderRD' SHA256: 6972e6ea6ece204edb2197ac92f699ad4e7c4b22cac23422dd816a2a9fc11ce3 Shader 'CubemapRoughnessShaderRD' SHA256: 083d83f629fde69ed8e4c482638f30df40144b9e05c8d8243317b1866ff97897 Shader 'TonemapShaderRD' SHA256: cd4d7156689aa910864bd876e2c439c8f2b49dd281df707f93145fec99b9bd67 Shader 'LuminanceReduceShaderRD' SHA256: f8db4abaa10d52974aef5a0443e69f8838eb05f4b4eee120c19ca3b0f8ef6353 Shader 'CubeToDpShaderRD' SHA256: 1ac5646aaebfed0f2fa62c6faa2f87981b8a5acf4748191482448d7fe516cc78 Shader 'BokehDofShaderRD' SHA256: 7ae11ecc5f3863fff19a28c9e41a9ba8016c53f54cc182daa15e6a1e080ce3c6 Shader 'SsaoDownsampleShaderRD' SHA256: 4db35346ca9c82436e993cde3466bee499a8f66f6d58e39d9bed78ee7d835e1f Shader 'SsaoShaderRD' SHA256: db7ede5f05458aaa5e03db5e62804bcd7e3b4b3eeb556e9eb027a958a96b6bd7 Shader 'SsaoImportanceMapShaderRD' SHA256: a2305d0dc9bed6da701e58b6d6a28bb2a664aedb9de4460c390b21a7f989e8eb Shader 'SsaoBlurShaderRD' SHA256: 9158ebb0e50201b64d1c1009317e7cf5de0792a82a4d06b393e5db5ec351b54f Shader 'SsaoInterleaveShaderRD' SHA256: e8d5eda46653d75788db067517688755630bcb40ff6da19c5f039db86df234c3 Shader 'RoughnessLimiterShaderRD' SHA256: fe30b9757ac0d3bbb53ab48bc7e0d48cc760f8212fc723c588c30dc9466539d4 Shader 'CubemapDownsamplerShaderRD' SHA256: 8184770cb634e8d5b7f7d4e801b864caa2972a7a98181a62c196ffa7b882edc3 Shader 'CubemapFilterShaderRD' SHA256: 5c27e7fe5c0fce79c147c7a8ae6984a0a7ae2dd90af07e35a4a34934698a91ca Shader 'SpecularMergeShaderRD' SHA256: f6dc7afac1a2715cc2f134daff1bddfd6b7ca715d655aca8cafa32798395dec2 Shader 'ScreenSpaceReflectionShaderRD' SHA256: d56b77268d5c239c3886ca85201d733c47ea7e6542969c5eb6d5350b483778ef Shader 'ScreenSpaceReflectionFilterShaderRD' SHA256: 4cd6afe65db3d97e059daa77702faf2a6e6cccb9c552c61af6055cb5be448b5f Shader 'ScreenSpaceReflectionScaleShaderRD' SHA256: 89c672345ad85d25a8e9cbc814aac4007cb5fcb29cb8262d0185abd23fdccb9f Shader 'SubsurfaceScatteringShaderRD' SHA256: c633c5beb82a7102c2849bec06732e4c6c98ebfc17b013f0844089d30beaef52 Shader 'ResolveShaderRD' SHA256: d0c9281d4f856dc19c5c6d14e0b256b28ca5455371db93744c010476572dbf25 Shader 'SortShaderRD' SHA256: 2192a5f1b643af9d53a2a609c7d2a715237ea16c083317a1611af42c6845b99c Shader 'BlitShaderRD' SHA256: 3c3477999355f834dd2a898d94df5b40ae97299bb1f5e9de2dd2b6caf46055f5 WASAPI: wFormatTag = 65534 WASAPI: nChannels = 8 WASAPI: nSamplesPerSec = 48000 WASAPI: nAvgBytesPerSec = 1536000 WASAPI: nBlockAlign = 32 WASAPI: wBitsPerSample = 32 WASAPI: cbSize = 22 WASAPI: detected 8 channels WASAPI: audio buffer frames: 1962 calculated latency: 44ms ERROR: Attempted to free invalid ID: 2100239007746 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) CORE API HASH: 12701809910231588744 EDITOR API HASH: 7657489752925521230 Using present mode: VK_PRESENT_MODE_FIFO_KHR Loaded builtin certs EditorSettings: Save OK! Using present mode: VK_PRESENT_MODE_FIFO_KHR Using present mode: VK_PRESENT_MODE_FIFO_KHR Using present mode: VK_PRESENT_MODE_FIFO_KHR EditorSettings: Save OK! ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: 1 RID allocations of type 'N17RendererStorageRD7TextureE' were leaked at exit. WARNING: 2 RIDs of type "Texture" were leaked. at: finalize (drivers\vulkan\rendering_device_vulkan.cpp:9055) StringName: 173 unclaimed string names at exit. ``` </details> Summary of tried combinations and result of it: | Distribution | MinGW version | GCC version | binutils version | Works | |--------------|------------------------|-------------|------------------|-------| | Fedora 34 | 8.0.0 | 10.3.1 | 2.34 | yes | | Fedora 35 | 9.0.0 | 11.2.1 | 2.37 | no | | Debian Sid | 8.0.0 | 10.2.1 | 2.37 | no | | OpenSuse TW | 9.0.0 | 9.2.0 | 2.33 | yes | | MSYS2 | 9.0.0.6346.6cc97775a-1 | 11.2.0 | 2.37 | no | | Mageia 8 | 8.0.0 | 10.2.1 | 2.34 | yes<sup>[1]</sup> | | MSYS2 | 9.0.0.6454.b4445ee52 | 11.2.0 | 2.38 | no | | MSYS2 | 10.0.0.r0.gaa08f56da-1 | 11.2.0 | 2.38 | no | [1] The icon from the executable was missing ### Steps to reproduce For a not working build: `scons platform=windows target=release_debug use_mingw=yes -j12` For a working build: `scons platform=windows target=debug use_mingw=yes -j12` ### Minimal reproduction project _No response_
1.0
Godot 4 compiled with MinGW is crashing with a white screen [GCC 11.x bug fixed in 11.3] - ### Godot version 4.0 custom build (bdcc8741e) ### System information Windows 10 21H1 ### Issue description The editor is crashing with a white screen on startup, with or without a project. I have been tried some combinations of compilers and options: MSVC with debug -> OK MSVC with release_debug -> OK MinGW with debug -> OK MinGW with release_debug -> Crash This issue has been introduced with the 99fe462452be44efa618e83ad9bbecd722ae6ecd commit. <details> <summary>Output of drmingw</summary> ``` godot.windows.opt.tools.64.exe caused an Access Violation at location 00007FF73A4C0000 DEP violation at location 00007FF73A4C0000. AddrPC Params 00007FF73A4C0000 00000215616FA200 0000003CE61FF5A8 0000003CE61FF5B0 00007FF77C559842 0000000000400000 00007FF77CA33F03 0000000000400000 godot.windows.opt.tools.64.exe!RenderingServerDefault::RenderingServerDefault [E:/path/core/os/thread.h @ 90] 88: _FORCE_INLINE_ ID get_id() const { return id; } 89: // get the ID of the caller thread > 90: _FORCE_INLINE_ static ID get_caller_id() { return caller_id; } 91: // get the ID of the main thread 92: _FORCE_INLINE_ static ID get_main_id() { return main_thread_id; } 00007FF77A4DE90B 0000021561579CC0 0000021560C1F690 00000215000000FF godot.windows.opt.tools.64.exe!Main::setup2 [E:/path/main/main.cpp @ 1590] 1588: /* Initialize Rendering Server */ 1589: > 1590: rendering_server = memnew(RenderingServerDefault(OS::get_singleton()->get_render_thread_mode() == OS::RENDER_SEPARATE_THREAD)); 1591: 1592: rendering_server->init(); 00007FF77A4E479B 0000021560C20C30 0000000000000001 000002155F450860 godot.windows.opt.tools.64.exe!Main::setup [E:/path/main/main.cpp @ 1388] 1386: 1387: if (p_second_phase) { > 1388: return setup2(); 1389: } 1390: 00007FF77A4C18A7 000002155F153296 0000003CE61FFD4C 0000000000000001 godot.windows.opt.tools.64.exe!widechar_main [E:/path/platform/windows/godot_windows.cpp @ 151] 149: TEST_MAIN_PARAM_OVERRIDE(argc, argv_utf8) 150: > 151: Error err = Main::setup(argv_utf8[0], argc - 1, &argv_utf8[1]); 152: 153: if (err != OK) { 00007FF77A4C19A0 0000000000000000 0000000000000047 00007FF7811164D8 godot.windows.opt.tools.64.exe!_main [E:/path/platform/windows/godot_windows.cpp @ 185] 183: } 184: > 185: result = widechar_main(argc, wc_argv); 186: 187: LocalFree(wc_argv); 00007FF77A4C13B1 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!__tmainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 321] 00007FF77A4C14E6 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!mainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 202] 00007FF8F8997034 0000000000000000 0000000000000000 0000000000000000 KERNEL32.DLL!BaseThreadInitThunk 00007FF8FA1C2651 0000000000000000 0000000000000000 0000000000000000 ntdll.dll!RtlUserThreadStart ``` </details> <details> <summary>Log of the crashing MinGW release_debug build</summary> ``` Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org Using "ICU / HarfBuzz / Graphite" text server... Vulkan API 1.2.162 ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157) WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files. Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154) Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080 - Vulkan multiview supported: max view count: 32 max instances: 134217727 - Vulkan subgroup: size: 32 stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV quad operations in all stages Using present mode: VK_PRESENT_MODE_FIFO_KHR Using "winink" pen tablet driver... ``` </details> <details> <summary>Log of the working MinGW debug build</summary> ``` Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org Using "ICU / HarfBuzz / Graphite" text server... Vulkan API 1.2.162 ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157) WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files. Objects - 1 Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080 at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154) Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080 - Vulkan multiview supported: max view count: 32 max instances: 134217727 - Vulkan subgroup: size: 32 stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV quad operations in all stages Using present mode: VK_PRESENT_MODE_FIFO_KHR Using "winink" pen tablet driver... Shader 'VoxelGiSdfShaderRD' SHA256: b6972a55781e21cfbac0dfc035381ee1f65e972c1f7e0bcc6004bfb74cc9dc9e Shader 'ParticlesShaderRD' SHA256: cbb6a5ce1933875a8e2df4f0cde4684286dedb6737cfbcfdeba745c978a2cc45 Shader 'ParticlesCopyShaderRD' SHA256: 6bd2ea6fca4c3a51f1e9bcd5068315e63820ec68fd8596e20480174c943bb9df Shader 'CanvasSdfShaderRD' SHA256: 692328d90ee7d404503d304e2464af3f3ac466ebd287762fd8883248f3e2a046 Shader 'SkeletonShaderRD' SHA256: 8f03a2ebc895d833120c3b2341c16e971fb1404dd3ae9e8030bc4bbf63e67fd5 Shader 'CanvasShaderRD' SHA256: a30a00a53b17f4e413dfb1fe109868063f2d67900aaeec66b0f50964108e216b Shader 'CanvasOcclusionShaderRD' SHA256: fd08d5c92b6537852fb4b8e5fba27f28518faad41a5c21ac9bd51bf49807824c Shader 'ClusterRenderShaderRD' SHA256: a2924afb04064a1dc42ed9d63721e7d5dd0466f245037ff85b5bef8bd251f515 Shader 'ClusterStoreShaderRD' SHA256: 64c437e6bebba9faa4766b524281310d934fd2a28f26e9a8a7834ca07f09d2dc Shader 'ClusterDebugShaderRD' SHA256: 0f66ac91d81e6f440ffa804d473579cb5d7ac135b375dce6d120186a65b54204 Shader 'SkyShaderRD' SHA256: dd0dbf614e400cc2f58360d292f85f19237fdd9746b9a46c6fd7c50fe8a2ae43 Shader 'VoxelGiShaderRD' SHA256: 51aa13996be8fb6071160d3ea73d752edcd67ef81e1aec1ec3623a13eee94d0a Shader 'VoxelGiDebugShaderRD' SHA256: 7fa6be49a459f8891a6e756165f2e9ca7f3ba261f53edf36c2041dc2acbf2a96 Shader 'SdfgiPreprocessShaderRD' SHA256: 056a9e6eaae9cc98093ceec6142bcb4582be0f3443e18402c1e25c293899b104 Shader 'SdfgiDirectLightShaderRD' SHA256: 0707e534eb11aeb7566a94913fe168d96b66896b440650f768f92ec6ed5cc842 Shader 'SdfgiIntegrateShaderRD' SHA256: deacba90b0c075f22f4000295f6a0b846957d2602b5d54b631c089792288e7df Shader 'GiShaderRD' SHA256: c1d65ee198e82b396759752f26b64ae98d7bb8691aed1bc07924617cf25872c5 Shader 'SdfgiDebugShaderRD' SHA256: 5e9f1e4d567be5c04dacec2ddbddb98638086a1fe7dc1d0260f3a58e93856a0e Shader 'SdfgiDebugProbesShaderRD' SHA256: a6a825191c0869482f513376305eb79fa16189e65068bf76a484a0e1eb602c1b Shader 'VolumetricFogShaderRD' SHA256: bc8fa35c7d92fd0353d5458af7d19438bf0c46f8c873b1b5e45f328f988f2d34 Shader 'SceneForwardClusteredShaderRD' SHA256: 300732714cbaf090328e341566d4d056b1bf83a2786d31de56778f6c929d1987 Shader 'CopyShaderRD' SHA256: 4ea51535b5cdcba7ece20ebd074657d05cbde47d1d0e87d3c84a2a1c28aa976f Shader 'CopyToFbShaderRD' SHA256: 6972e6ea6ece204edb2197ac92f699ad4e7c4b22cac23422dd816a2a9fc11ce3 Shader 'CubemapRoughnessShaderRD' SHA256: 083d83f629fde69ed8e4c482638f30df40144b9e05c8d8243317b1866ff97897 Shader 'TonemapShaderRD' SHA256: cd4d7156689aa910864bd876e2c439c8f2b49dd281df707f93145fec99b9bd67 Shader 'LuminanceReduceShaderRD' SHA256: f8db4abaa10d52974aef5a0443e69f8838eb05f4b4eee120c19ca3b0f8ef6353 Shader 'CubeToDpShaderRD' SHA256: 1ac5646aaebfed0f2fa62c6faa2f87981b8a5acf4748191482448d7fe516cc78 Shader 'BokehDofShaderRD' SHA256: 7ae11ecc5f3863fff19a28c9e41a9ba8016c53f54cc182daa15e6a1e080ce3c6 Shader 'SsaoDownsampleShaderRD' SHA256: 4db35346ca9c82436e993cde3466bee499a8f66f6d58e39d9bed78ee7d835e1f Shader 'SsaoShaderRD' SHA256: db7ede5f05458aaa5e03db5e62804bcd7e3b4b3eeb556e9eb027a958a96b6bd7 Shader 'SsaoImportanceMapShaderRD' SHA256: a2305d0dc9bed6da701e58b6d6a28bb2a664aedb9de4460c390b21a7f989e8eb Shader 'SsaoBlurShaderRD' SHA256: 9158ebb0e50201b64d1c1009317e7cf5de0792a82a4d06b393e5db5ec351b54f Shader 'SsaoInterleaveShaderRD' SHA256: e8d5eda46653d75788db067517688755630bcb40ff6da19c5f039db86df234c3 Shader 'RoughnessLimiterShaderRD' SHA256: fe30b9757ac0d3bbb53ab48bc7e0d48cc760f8212fc723c588c30dc9466539d4 Shader 'CubemapDownsamplerShaderRD' SHA256: 8184770cb634e8d5b7f7d4e801b864caa2972a7a98181a62c196ffa7b882edc3 Shader 'CubemapFilterShaderRD' SHA256: 5c27e7fe5c0fce79c147c7a8ae6984a0a7ae2dd90af07e35a4a34934698a91ca Shader 'SpecularMergeShaderRD' SHA256: f6dc7afac1a2715cc2f134daff1bddfd6b7ca715d655aca8cafa32798395dec2 Shader 'ScreenSpaceReflectionShaderRD' SHA256: d56b77268d5c239c3886ca85201d733c47ea7e6542969c5eb6d5350b483778ef Shader 'ScreenSpaceReflectionFilterShaderRD' SHA256: 4cd6afe65db3d97e059daa77702faf2a6e6cccb9c552c61af6055cb5be448b5f Shader 'ScreenSpaceReflectionScaleShaderRD' SHA256: 89c672345ad85d25a8e9cbc814aac4007cb5fcb29cb8262d0185abd23fdccb9f Shader 'SubsurfaceScatteringShaderRD' SHA256: c633c5beb82a7102c2849bec06732e4c6c98ebfc17b013f0844089d30beaef52 Shader 'ResolveShaderRD' SHA256: d0c9281d4f856dc19c5c6d14e0b256b28ca5455371db93744c010476572dbf25 Shader 'SortShaderRD' SHA256: 2192a5f1b643af9d53a2a609c7d2a715237ea16c083317a1611af42c6845b99c Shader 'BlitShaderRD' SHA256: 3c3477999355f834dd2a898d94df5b40ae97299bb1f5e9de2dd2b6caf46055f5 WASAPI: wFormatTag = 65534 WASAPI: nChannels = 8 WASAPI: nSamplesPerSec = 48000 WASAPI: nAvgBytesPerSec = 1536000 WASAPI: nBlockAlign = 32 WASAPI: wBitsPerSample = 32 WASAPI: cbSize = 22 WASAPI: detected 8 channels WASAPI: audio buffer frames: 1962 calculated latency: 44ms ERROR: Attempted to free invalid ID: 2100239007746 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) CORE API HASH: 12701809910231588744 EDITOR API HASH: 7657489752925521230 Using present mode: VK_PRESENT_MODE_FIFO_KHR Loaded builtin certs EditorSettings: Save OK! Using present mode: VK_PRESENT_MODE_FIFO_KHR Using present mode: VK_PRESENT_MODE_FIFO_KHR Using present mode: VK_PRESENT_MODE_FIFO_KHR EditorSettings: Save OK! ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: Attempted to free invalid ID: 0 at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313) ERROR: 1 RID allocations of type 'N17RendererStorageRD7TextureE' were leaked at exit. WARNING: 2 RIDs of type "Texture" were leaked. at: finalize (drivers\vulkan\rendering_device_vulkan.cpp:9055) StringName: 173 unclaimed string names at exit. ``` </details> Summary of tried combinations and result of it: | Distribution | MinGW version | GCC version | binutils version | Works | |--------------|------------------------|-------------|------------------|-------| | Fedora 34 | 8.0.0 | 10.3.1 | 2.34 | yes | | Fedora 35 | 9.0.0 | 11.2.1 | 2.37 | no | | Debian Sid | 8.0.0 | 10.2.1 | 2.37 | no | | OpenSuse TW | 9.0.0 | 9.2.0 | 2.33 | yes | | MSYS2 | 9.0.0.6346.6cc97775a-1 | 11.2.0 | 2.37 | no | | Mageia 8 | 8.0.0 | 10.2.1 | 2.34 | yes<sup>[1]</sup> | | MSYS2 | 9.0.0.6454.b4445ee52 | 11.2.0 | 2.38 | no | | MSYS2 | 10.0.0.r0.gaa08f56da-1 | 11.2.0 | 2.38 | no | [1] The icon from the executable was missing ### Steps to reproduce For a not working build: `scons platform=windows target=release_debug use_mingw=yes -j12` For a working build: `scons platform=windows target=debug use_mingw=yes -j12` ### Minimal reproduction project _No response_
non_process
godot compiled with mingw is crashing with a white screen godot version custom build system information windows issue description the editor is crashing with a white screen on startup with or without a project i have been tried some combinations of compilers and options msvc with debug ok msvc with release debug ok mingw with debug ok mingw with release debug crash this issue has been introduced with the commit output of drmingw godot windows opt tools exe caused an access violation at location dep violation at location addrpc params godot windows opt tools exe renderingserverdefault renderingserverdefault force inline id get id const return id get the id of the caller thread force inline static id get caller id return caller id get the id of the main thread force inline static id get main id return main thread id godot windows opt tools exe main initialize rendering server rendering server memnew renderingserverdefault os get singleton get render thread mode os render separate thread rendering server init godot windows opt tools exe main setup if p second phase return godot windows opt tools exe widechar main test main param override argc argv error err main setup argv argc argv if err ok godot windows opt tools exe main result widechar main argc wc argv localfree wc argv godot windows opt tools exe tmaincrtstartup godot windows opt tools exe maincrtstartup dll basethreadinitthunk ntdll dll rtluserthreadstart log of the crashing mingw release debug build godot engine dev custom build using icu harfbuzz graphite text server vulkan api error general message id number message id name loader message loader get json failed to open json file c programdata gog com galaxy redists overlay injected galaxy overlay vklayer json objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp warning general message id number message id name loader message readdatafilesinregistry registry lookup failed to get layer manifest files objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp using vulkan device nvidia nvidia geforce rtx vulkan multiview supported max view count max instances vulkan subgroup size stages stage vertex stage tessellation control stage tessellation evaluation stage geometry stage fragment stage compute stage raygen khr stage any hit khr stage closest hit khr stage miss khr stage intersection khr stage callable khr stage task nv stage mesh nv supported ops feature basic feature vote feature arithmetic feature ballot feature shuffle feature shuffle relative feature clustered feature quad feature partitioned nv quad operations in all stages using present mode vk present mode fifo khr using winink pen tablet driver log of the working mingw debug build godot engine dev custom build using icu harfbuzz graphite text server vulkan api error general message id number message id name loader message loader get json failed to open json file c programdata gog com galaxy redists overlay injected galaxy overlay vklayer json objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp warning general message id number message id name loader message readdatafilesinregistry registry lookup failed to get layer manifest files objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp using vulkan device nvidia nvidia geforce rtx vulkan multiview supported max view count max instances vulkan subgroup size stages stage vertex stage tessellation control stage tessellation evaluation stage geometry stage fragment stage compute stage raygen khr stage any hit khr stage closest hit khr stage miss khr stage intersection khr stage callable khr stage task nv stage mesh nv supported ops feature basic feature vote feature arithmetic feature ballot feature shuffle feature shuffle relative feature clustered feature quad feature partitioned nv quad operations in all stages using present mode vk present mode fifo khr using winink pen tablet driver shader voxelgisdfshaderrd shader particlesshaderrd shader particlescopyshaderrd shader canvassdfshaderrd shader skeletonshaderrd shader canvasshaderrd shader canvasocclusionshaderrd shader clusterrendershaderrd shader clusterstoreshaderrd shader clusterdebugshaderrd shader skyshaderrd shader voxelgishaderrd shader voxelgidebugshaderrd shader sdfgipreprocessshaderrd shader sdfgidirectlightshaderrd shader sdfgiintegrateshaderrd shader gishaderrd shader sdfgidebugshaderrd shader sdfgidebugprobesshaderrd shader volumetricfogshaderrd shader sceneforwardclusteredshaderrd shader copyshaderrd shader copytofbshaderrd shader cubemaproughnessshaderrd shader tonemapshaderrd shader luminancereduceshaderrd shader cubetodpshaderrd shader bokehdofshaderrd shader ssaodownsampleshaderrd shader ssaoshaderrd shader ssaoimportancemapshaderrd shader ssaoblurshaderrd shader ssaointerleaveshaderrd shader roughnesslimitershaderrd shader cubemapdownsamplershaderrd shader cubemapfiltershaderrd shader specularmergeshaderrd shader screenspacereflectionshaderrd shader screenspacereflectionfiltershaderrd shader screenspacereflectionscaleshaderrd shader subsurfacescatteringshaderrd shader resolveshaderrd shader sortshaderrd shader blitshaderrd wasapi wformattag wasapi nchannels wasapi nsamplespersec wasapi navgbytespersec wasapi nblockalign wasapi wbitspersample wasapi cbsize wasapi detected channels wasapi audio buffer frames calculated latency error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp core api hash editor api hash using present mode vk present mode fifo khr loaded builtin certs editorsettings save ok using present mode vk present mode fifo khr using present mode vk present mode fifo khr using present mode vk present mode fifo khr editorsettings save ok error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error rid allocations of type were leaked at exit warning rids of type texture were leaked at finalize drivers vulkan rendering device vulkan cpp stringname unclaimed string names at exit summary of tried combinations and result of it distribution mingw version gcc version binutils version works fedora yes fedora no debian sid no opensuse tw yes no mageia yes no no the icon from the executable was missing steps to reproduce for a not working build scons platform windows target release debug use mingw yes for a working build scons platform windows target debug use mingw yes minimal reproduction project no response
0
17,578
23,389,698,328
IssuesEvent
2022-08-11 16:35:22
vectordotdev/vector
https://api.github.com/repos/vectordotdev/vector
closed
New `javascript` transform
needs: approval domain: transforms type: feature domain: processing
Recently Fabrice Bellard, creator of QEMU and FFMPEG, released [QuickJS](https://bellard.org/quickjs/) embeddable JavaScript engine. It is MIT licensed and its compiled size is just a few hundred kilobytes. So it might be a good idea to add `javascript` transform, similar to `lua` transform.
1.0
New `javascript` transform - Recently Fabrice Bellard, creator of QEMU and FFMPEG, released [QuickJS](https://bellard.org/quickjs/) embeddable JavaScript engine. It is MIT licensed and its compiled size is just a few hundred kilobytes. So it might be a good idea to add `javascript` transform, similar to `lua` transform.
process
new javascript transform recently fabrice bellard creator of qemu and ffmpeg released embeddable javascript engine it is mit licensed and its compiled size is just a few hundred kilobytes so it might be a good idea to add javascript transform similar to lua transform
1
18,132
24,171,033,799
IssuesEvent
2022-09-22 19:14:25
ORNL-AMO/AMO-Tools-Suite
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite
closed
Gas Flue Gas Edit
bug Process Heating
Issue overview -------------- GasFlueGasMaterial.cpp line 110 Change to... const double combAirMoisture = combAirMoisturePerc <= 0.009 ? 0 : ((combAirMoisturePerc - 0.009) * (stoichAir * (1 + exsAir)) * 0.0763);
1.0
Gas Flue Gas Edit - Issue overview -------------- GasFlueGasMaterial.cpp line 110 Change to... const double combAirMoisture = combAirMoisturePerc <= 0.009 ? 0 : ((combAirMoisturePerc - 0.009) * (stoichAir * (1 + exsAir)) * 0.0763);
process
gas flue gas edit issue overview gasfluegasmaterial cpp line change to const double combairmoisture combairmoistureperc combairmoistureperc stoichair exsair
1
19,404
25,544,499,540
IssuesEvent
2022-11-29 17:38:00
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process: spawn incorrect error
child_process
### Version v18.5.0 ### Platform 5.4.0-88-generic #99-Ubuntu SMP Thu Sep 23 17:29:00 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux ### Subsystem _No response_ ### What steps will reproduce the bug? ```js const spawn = require('child_process').spawn; const cwd = magicFunction() // function return failed path: '/tmp1' - when debugging is not immediately visible const cp = spawn('ls', [], { cwd }); cp.stdout.on('data', function(data) { console.log(data.toString()); }); cp.stderr.on('data', function(data) { console.log(data.toString()); }); cp.on('exit', function(code) {}); ``` ### How often does it reproduce? Is there a required condition? _No response_ ### What is the expected behavior? Error: spawn cwd ENOENT ### What do you see instead? Error: spawn ls ENOENT ### Additional information Hi, when calling system commands error is misleading, literally, spawn can't find ls, actually the problem is in cwd. It seemed that the problem is with the first argument `ls` --> `/bin/bash` and two argument `[]` --> `['-c', 'ls']` or systems environment OS PATH but not cwd
1.0
child_process: spawn incorrect error - ### Version v18.5.0 ### Platform 5.4.0-88-generic #99-Ubuntu SMP Thu Sep 23 17:29:00 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux ### Subsystem _No response_ ### What steps will reproduce the bug? ```js const spawn = require('child_process').spawn; const cwd = magicFunction() // function return failed path: '/tmp1' - when debugging is not immediately visible const cp = spawn('ls', [], { cwd }); cp.stdout.on('data', function(data) { console.log(data.toString()); }); cp.stderr.on('data', function(data) { console.log(data.toString()); }); cp.on('exit', function(code) {}); ``` ### How often does it reproduce? Is there a required condition? _No response_ ### What is the expected behavior? Error: spawn cwd ENOENT ### What do you see instead? Error: spawn ls ENOENT ### Additional information Hi, when calling system commands error is misleading, literally, spawn can't find ls, actually the problem is in cwd. It seemed that the problem is with the first argument `ls` --> `/bin/bash` and two argument `[]` --> `['-c', 'ls']` or systems environment OS PATH but not cwd
process
child process spawn incorrect error version platform generic ubuntu smp thu sep utc gnu linux subsystem no response what steps will reproduce the bug js const spawn require child process spawn const cwd magicfunction function return failed path when debugging is not immediately visible const cp spawn ls cwd cp stdout on data function data console log data tostring cp stderr on data function data console log data tostring cp on exit function code how often does it reproduce is there a required condition no response what is the expected behavior error spawn cwd enoent what do you see instead error spawn ls enoent additional information hi when calling system commands error is misleading literally spawn can t find ls actually the problem is in cwd it seemed that the problem is with the first argument ls bin bash and two argument or systems environment os path but not cwd
1
14,008
16,814,616,075
IssuesEvent
2021-06-17 05:24:05
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
GitLab source control integration?
Pri2 automation/svc awaiting-product-team-response cxp process-automation/subsvc product-question triaged
[ Hi, Is there a way to use GitLab for SCI with Azure Automation? ] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
GitLab source control integration? - [ Hi, Is there a way to use GitLab for SCI with Azure Automation? ] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
gitlab source control integration hi is there a way to use gitlab for sci with azure automation document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
5,328
8,142,119,958
IssuesEvent
2018-08-21 06:16:49
Madek/madek
https://api.github.com/repos/Madek/madek
opened
Tests Batch Editing
Batch process
- I would expect the metatags to be sorted alphabetically ![bildschirmfoto 2018-08-21 um 07 51 41](https://user-images.githubusercontent.com/3356016/44382899-116d3280-a517-11e8-926b-a9e50c2dfba6.png) - If I press "anwenden", then I do not see any kind of notice on the tile itself. I would like to see a change in colour or the button "anwenden" should vanish. ![bildschirmfoto 2018-08-21 um 07 54 45](https://user-images.githubusercontent.com/3356016/44383044-93f5f200-a517-11e8-9dd8-8d37aaf004dc.png) - Can you make the popup bigger, so the Derivates is not on a single line? ![bildschirmfoto 2018-08-21 um 08 03 46](https://user-images.githubusercontent.com/3356016/44383376-d966ef00-a518-11e8-9ff4-7f486e8a1158.png) - It would be nice, if the amount of updated media entries would be added to the green message (Erfolgsmeldung) and if the message would not vanish ![bildschirmfoto 2018-08-21 um 08 00 22](https://user-images.githubusercontent.com/3356016/44383231-547bd580-a518-11e8-90be-227af1e856d9.png) - If the filter is opened, it is hard to differentiate which is the filter and which are the fields to pick. ![bildschirmfoto 2018-08-21 um 08 10 20](https://user-images.githubusercontent.com/3356016/44383651-acffa280-a519-11e8-8c8e-1d8a6ecdceba.png) - Please write "Medien Einträgen" together "Medieneinträgen" ![bildschirmfoto 2018-08-21 um 08 15 25](https://user-images.githubusercontent.com/3356016/44383837-6b232c00-a51a-11e8-8b4a-937e547752d9.png)
1.0
Tests Batch Editing - - I would expect the metatags to be sorted alphabetically ![bildschirmfoto 2018-08-21 um 07 51 41](https://user-images.githubusercontent.com/3356016/44382899-116d3280-a517-11e8-926b-a9e50c2dfba6.png) - If I press "anwenden", then I do not see any kind of notice on the tile itself. I would like to see a change in colour or the button "anwenden" should vanish. ![bildschirmfoto 2018-08-21 um 07 54 45](https://user-images.githubusercontent.com/3356016/44383044-93f5f200-a517-11e8-9dd8-8d37aaf004dc.png) - Can you make the popup bigger, so the Derivates is not on a single line? ![bildschirmfoto 2018-08-21 um 08 03 46](https://user-images.githubusercontent.com/3356016/44383376-d966ef00-a518-11e8-9ff4-7f486e8a1158.png) - It would be nice, if the amount of updated media entries would be added to the green message (Erfolgsmeldung) and if the message would not vanish ![bildschirmfoto 2018-08-21 um 08 00 22](https://user-images.githubusercontent.com/3356016/44383231-547bd580-a518-11e8-90be-227af1e856d9.png) - If the filter is opened, it is hard to differentiate which is the filter and which are the fields to pick. ![bildschirmfoto 2018-08-21 um 08 10 20](https://user-images.githubusercontent.com/3356016/44383651-acffa280-a519-11e8-8c8e-1d8a6ecdceba.png) - Please write "Medien Einträgen" together "Medieneinträgen" ![bildschirmfoto 2018-08-21 um 08 15 25](https://user-images.githubusercontent.com/3356016/44383837-6b232c00-a51a-11e8-8b4a-937e547752d9.png)
process
tests batch editing i would expect the metatags to be sorted alphabetically if i press anwenden then i do not see any kind of notice on the tile itself i would like to see a change in colour or the button anwenden should vanish can you make the popup bigger so the derivates is not on a single line it would be nice if the amount of updated media entries would be added to the green message erfolgsmeldung and if the message would not vanish if the filter is opened it is hard to differentiate which is the filter and which are the fields to pick please write medien einträgen together medieneinträgen
1
495,202
14,273,523,211
IssuesEvent
2020-11-21 22:07:30
tqZK/harc_game_web
https://api.github.com/repos/tqZK/harc_game_web
closed
Add `category` field to Task model
backend base-version good first issue priority
Also, in view with list of tasks, divide them into categories
1.0
Add `category` field to Task model - Also, in view with list of tasks, divide them into categories
non_process
add category field to task model also in view with list of tasks divide them into categories
0
9,751
12,737,044,163
IssuesEvent
2020-06-25 18:01:12
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
PreProcessor files for easy Progmatic PreProcessor Definitions
area-System.Diagnostics.Process untriaged
so, this is a feature request, that might as well be for MSBuild as a whole. but no clue where to post that so.... The Problem: You have a library that you want to make possible to disable or remove certain features that you wouldn't need in a safe and practical way And you do this via PreProcessor Directives but if you wish to take it up a notch and make large libraries with multiple disablable portions you quickly run into the issue of preprocessors taking up a large portion of every file and being mostly unreadable or you needing to define multiple PreProcessors combine this with open source and having the end-user define what they need and what they don't you have an almost impossible problem. Currently only remedied via custom or cumbersome PreCompilers or Pre-Compile Processors which just add to the complexity of a project. The Suggestion: a new file type for MSBuild *.ppf (Pre Processor File) a file that just consists of preprocessor directives, that will get processed and the resulting state will then constitute the global preprocessing state, IE if you have something like this. [MyPreprocessor.pff] ``` #if Custom_Disable_Directive1 #define Disabled_Internal_Something_Directive #if Conflicting_Directive #error X depends on Y, You must choose #endif #endif ``` the code above would be run before processing all the other files and all defined processors added to the global store to act across all cs files.
1.0
PreProcessor files for easy Progmatic PreProcessor Definitions - so, this is a feature request, that might as well be for MSBuild as a whole. but no clue where to post that so.... The Problem: You have a library that you want to make possible to disable or remove certain features that you wouldn't need in a safe and practical way And you do this via PreProcessor Directives but if you wish to take it up a notch and make large libraries with multiple disablable portions you quickly run into the issue of preprocessors taking up a large portion of every file and being mostly unreadable or you needing to define multiple PreProcessors combine this with open source and having the end-user define what they need and what they don't you have an almost impossible problem. Currently only remedied via custom or cumbersome PreCompilers or Pre-Compile Processors which just add to the complexity of a project. The Suggestion: a new file type for MSBuild *.ppf (Pre Processor File) a file that just consists of preprocessor directives, that will get processed and the resulting state will then constitute the global preprocessing state, IE if you have something like this. [MyPreprocessor.pff] ``` #if Custom_Disable_Directive1 #define Disabled_Internal_Something_Directive #if Conflicting_Directive #error X depends on Y, You must choose #endif #endif ``` the code above would be run before processing all the other files and all defined processors added to the global store to act across all cs files.
process
preprocessor files for easy progmatic preprocessor definitions so this is a feature request that might as well be for msbuild as a whole but no clue where to post that so the problem you have a library that you want to make possible to disable or remove certain features that you wouldn t need in a safe and practical way and you do this via preprocessor directives but if you wish to take it up a notch and make large libraries with multiple disablable portions you quickly run into the issue of preprocessors taking up a large portion of every file and being mostly unreadable or you needing to define multiple preprocessors combine this with open source and having the end user define what they need and what they don t you have an almost impossible problem currently only remedied via custom or cumbersome precompilers or pre compile processors which just add to the complexity of a project the suggestion a new file type for msbuild ppf pre processor file a file that just consists of preprocessor directives that will get processed and the resulting state will then constitute the global preprocessing state ie if you have something like this if custom disable define disabled internal something directive if conflicting directive error x depends on y you must choose endif endif the code above would be run before processing all the other files and all defined processors added to the global store to act across all cs files
1
131,746
10,708,674,656
IssuesEvent
2019-10-24 20:14:15
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Fix iOS text field tests
a: tests ☸ platform-web
There are few tests in `text_field_test.dart` that cover iOS-specific behavior. Those are currently failing.
1.0
Fix iOS text field tests - There are few tests in `text_field_test.dart` that cover iOS-specific behavior. Those are currently failing.
non_process
fix ios text field tests there are few tests in text field test dart that cover ios specific behavior those are currently failing
0
336,282
10,179,729,915
IssuesEvent
2019-08-09 08:32:25
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
closed
Re-sort content review queue by most recently updated
component: reviewer tools priority: p3
Currently the content review queue is sorted by "oldest to most recent" updates. We'd like to flip it to be sorted from "most recent to oldest" updates for the August 8 push.
1.0
Re-sort content review queue by most recently updated - Currently the content review queue is sorted by "oldest to most recent" updates. We'd like to flip it to be sorted from "most recent to oldest" updates for the August 8 push.
non_process
re sort content review queue by most recently updated currently the content review queue is sorted by oldest to most recent updates we d like to flip it to be sorted from most recent to oldest updates for the august push
0
14,047
16,851,695,381
IssuesEvent
2021-06-20 16:38:45
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
segfault in denoise (profiled) / nlmeans
bug: pending reproduce: random scope: image processing
**Describe the bug/issue** segfault (backtrace at https://pastebin.com/skhGKeh4) while editing raw from https://discuss.pixls.us/t/hockey-game-under-lights-20k-iso-on-m4-3/25502. Analysis on IRC: ``` [10:10] <@johnny_bit> darktable/src/common/nlmeans_core.c:694 caused the app to go into sigsegv ... [10:12] <@johnny_bit> it looks like the index i might go out of range for col_sums array ... [10:42] <@johnny_bit> You can raise an issue with that stacktrace and ping ralfbrown on it because it's part of the code he's now most "accustomed" with, adding the info we've managed to "guess" here :) ``` **To Reproduce** Edit settings of _denoise (profiled)_ in NL-means mode (does not happen immediately, probably hard to reproduce) **Expected behavior** No crash **Which commit introduced the error** I don't know, I hardly ever use NL-means. **Platform** * darktable version : 864e22daf * OS : Linux - kernel 5.11.0-18 * Linux - Distro : KUbuntu 21.04 * Memory : 64 GM * Graphics card : Nvidia 1060/6GB * Graphics driver : 460.80-0ubuntu0.21.04.2 * OpenCL installed : yes * OpenCL activated : yes * Xorg : xserver-xorg 1:7.7+22ubuntu1; xserver-xorg-core 2:1.20.11-1ubuntu1 * Desktop : KDE * GTK+ : 3.24.25-1ubuntu4.1 * gcc : 10.3.0 * cflags : * CMAKE_BUILD_TYPE : RelWithDebInfo
1.0
segfault in denoise (profiled) / nlmeans - **Describe the bug/issue** segfault (backtrace at https://pastebin.com/skhGKeh4) while editing raw from https://discuss.pixls.us/t/hockey-game-under-lights-20k-iso-on-m4-3/25502. Analysis on IRC: ``` [10:10] <@johnny_bit> darktable/src/common/nlmeans_core.c:694 caused the app to go into sigsegv ... [10:12] <@johnny_bit> it looks like the index i might go out of range for col_sums array ... [10:42] <@johnny_bit> You can raise an issue with that stacktrace and ping ralfbrown on it because it's part of the code he's now most "accustomed" with, adding the info we've managed to "guess" here :) ``` **To Reproduce** Edit settings of _denoise (profiled)_ in NL-means mode (does not happen immediately, probably hard to reproduce) **Expected behavior** No crash **Which commit introduced the error** I don't know, I hardly ever use NL-means. **Platform** * darktable version : 864e22daf * OS : Linux - kernel 5.11.0-18 * Linux - Distro : KUbuntu 21.04 * Memory : 64 GM * Graphics card : Nvidia 1060/6GB * Graphics driver : 460.80-0ubuntu0.21.04.2 * OpenCL installed : yes * OpenCL activated : yes * Xorg : xserver-xorg 1:7.7+22ubuntu1; xserver-xorg-core 2:1.20.11-1ubuntu1 * Desktop : KDE * GTK+ : 3.24.25-1ubuntu4.1 * gcc : 10.3.0 * cflags : * CMAKE_BUILD_TYPE : RelWithDebInfo
process
segfault in denoise profiled nlmeans describe the bug issue segfault backtrace at while editing raw from analysis on irc darktable src common nlmeans core c caused the app to go into sigsegv it looks like the index i might go out of range for col sums array you can raise an issue with that stacktrace and ping ralfbrown on it because it s part of the code he s now most accustomed with adding the info we ve managed to guess here to reproduce edit settings of denoise profiled in nl means mode does not happen immediately probably hard to reproduce expected behavior no crash which commit introduced the error i don t know i hardly ever use nl means platform darktable version os linux kernel linux distro kubuntu memory gm graphics card nvidia graphics driver opencl installed yes opencl activated yes xorg xserver xorg xserver xorg core desktop kde gtk gcc cflags cmake build type relwithdebinfo
1
10,724
8,696,995,781
IssuesEvent
2018-12-04 19:02:08
webhintio/hint
https://api.github.com/repos/webhintio/hint
closed
Provide more contribution related details
area:infrastructure
- [x] Add `ISSUE_TEMPLATE.md`. - [x] Add `PULL_REQUEST_TEMPLATE.md`. - [x] Add [`CODEOWNERS` file](https://github.com/blog/2392-introducing-code-owners) (see: #382). - [ ] Update `CONTRIBUTING.md` with more information. - [ ] Add [`SUPPORT` file](https://github.com/blog/2400-support-file-support).
1.0
Provide more contribution related details - - [x] Add `ISSUE_TEMPLATE.md`. - [x] Add `PULL_REQUEST_TEMPLATE.md`. - [x] Add [`CODEOWNERS` file](https://github.com/blog/2392-introducing-code-owners) (see: #382). - [ ] Update `CONTRIBUTING.md` with more information. - [ ] Add [`SUPPORT` file](https://github.com/blog/2400-support-file-support).
non_process
provide more contribution related details add issue template md add pull request template md add see update contributing md with more information add
0
106,507
23,243,713,552
IssuesEvent
2022-08-03 17:57:23
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
go-to-protobuf: type FooSpec has both field and method named Size
kind/bug sig/api-machinery area/code-generation lifecycle/rotten
**What happened**: go-to-protobuf generates `Size()` func which breaks the compilation when the type has `Size` field. **What you expected to happen**: go-to-protobuf generation to do not break the compilation. **How to reproduce it (as minimally and precisely as possible)**: 1. Assume that your type has a field named `Size`. I simple add this type to the `k8s.io/sample-apiserver` types.go. ```go type FooSpec struct { Size string `json:"size" protobuf:"bytes,1,opt,name=size"` } ``` 2. Run go-to-protobuf ``` $ go-to-protobuf \ --output-base="${GOPATH}/src" \ --apimachinery-packages='-k8s.io/apimachinery/pkg/util/intstr,-k8s.io/apimachinery/pkg/api/resource,-k8s.io/apimachinery/pkg/runtime/schema,-k8s.io/apimachinery/pkg/runtime,-k8s.io/apimachinery/pkg/apis/meta/v1,-k8s.io/apimachinery/pkg/apis/meta/v1beta1,-k8s.io/api/core/v1,-k8s.io/api/rbac/v1' \ --go-header-file=./hack/custom-boilerplate.go.txt \ --proto-import=./third_party/protobuf \ --proto-import=./vendor \ --packages="k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1" ``` 3. Check generated.pb.go and ensure it contains ```go // ... func (m *FooSpec) Size() (n int) { if m == nil { return 0 } var l int _ = l l = len(m.Size_) n += 1 + l + sovGenerated(uint64(l)) return n } // ... ``` 4. Ensure that the compilation fails ``` $ GOOS=linux go build . # k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1 pkg/apis/wardle/v1alpha1/generated.pb.go:563:12: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:564:18: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:565:49: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size pkg/apis/wardle/v1alpha1/generated.pb.go:678:11: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:769:36: this.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:1522:5: m.Size_ undefined (type *FooSpec has no field or method Size_) ``` and note the root cause ``` pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size ``` **Anything else we need to know?**: k8s.io/code-generator version - v1.16.0 **Environment**: - Kubernetes version (use `kubectl version`): - Cloud provider or hardware configuration: - OS (e.g: `cat /etc/os-release`): - Kernel (e.g. `uname -a`): - Install tools: - Network plugin and version (if this is a network-related bug): - Others:
1.0
go-to-protobuf: type FooSpec has both field and method named Size - **What happened**: go-to-protobuf generates `Size()` func which breaks the compilation when the type has `Size` field. **What you expected to happen**: go-to-protobuf generation to do not break the compilation. **How to reproduce it (as minimally and precisely as possible)**: 1. Assume that your type has a field named `Size`. I simple add this type to the `k8s.io/sample-apiserver` types.go. ```go type FooSpec struct { Size string `json:"size" protobuf:"bytes,1,opt,name=size"` } ``` 2. Run go-to-protobuf ``` $ go-to-protobuf \ --output-base="${GOPATH}/src" \ --apimachinery-packages='-k8s.io/apimachinery/pkg/util/intstr,-k8s.io/apimachinery/pkg/api/resource,-k8s.io/apimachinery/pkg/runtime/schema,-k8s.io/apimachinery/pkg/runtime,-k8s.io/apimachinery/pkg/apis/meta/v1,-k8s.io/apimachinery/pkg/apis/meta/v1beta1,-k8s.io/api/core/v1,-k8s.io/api/rbac/v1' \ --go-header-file=./hack/custom-boilerplate.go.txt \ --proto-import=./third_party/protobuf \ --proto-import=./vendor \ --packages="k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1" ``` 3. Check generated.pb.go and ensure it contains ```go // ... func (m *FooSpec) Size() (n int) { if m == nil { return 0 } var l int _ = l l = len(m.Size_) n += 1 + l + sovGenerated(uint64(l)) return n } // ... ``` 4. Ensure that the compilation fails ``` $ GOOS=linux go build . # k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1 pkg/apis/wardle/v1alpha1/generated.pb.go:563:12: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:564:18: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:565:49: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size pkg/apis/wardle/v1alpha1/generated.pb.go:678:11: m.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:769:36: this.Size_ undefined (type *FooSpec has no field or method Size_) pkg/apis/wardle/v1alpha1/generated.pb.go:1522:5: m.Size_ undefined (type *FooSpec has no field or method Size_) ``` and note the root cause ``` pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size ``` **Anything else we need to know?**: k8s.io/code-generator version - v1.16.0 **Environment**: - Kubernetes version (use `kubectl version`): - Cloud provider or hardware configuration: - OS (e.g: `cat /etc/os-release`): - Kernel (e.g. `uname -a`): - Install tools: - Network plugin and version (if this is a network-related bug): - Others:
non_process
go to protobuf type foospec has both field and method named size what happened go to protobuf generates size func which breaks the compilation when the type has size field what you expected to happen go to protobuf generation to do not break the compilation how to reproduce it as minimally and precisely as possible assume that your type has a field named size i simple add this type to the io sample apiserver types go go type foospec struct size string json size protobuf bytes opt name size run go to protobuf go to protobuf output base gopath src apimachinery packages io apimachinery pkg util intstr io apimachinery pkg api resource io apimachinery pkg runtime schema io apimachinery pkg runtime io apimachinery pkg apis meta io apimachinery pkg apis meta io api core io api rbac go header file hack custom boilerplate go txt proto import third party protobuf proto import vendor packages io sample apiserver pkg apis wardle check generated pb go and ensure it contains go func m foospec size n int if m nil return var l int l l len m size n l sovgenerated l return n ensure that the compilation fails goos linux go build io sample apiserver pkg apis wardle pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go type foospec has both field and method named size pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go this size undefined type foospec has no field or method size pkg apis wardle generated pb go m size undefined type foospec has no field or method size and note the root cause pkg apis wardle generated pb go type foospec has both field and method named size anything else we need to know io code generator version environment kubernetes version use kubectl version cloud provider or hardware configuration os e g cat etc os release kernel e g uname a install tools network plugin and version if this is a network related bug others
0
10,977
13,781,068,315
IssuesEvent
2020-10-08 15:41:47
w3c/webauthn
https://api.github.com/repos/w3c/webauthn
opened
Update IANA Registry with Apple attestation format
type:process
The [IANA Registry](https://www.iana.org/assignments/webauthn/webauthn.xhtml) should be updated to include Apple's new attestation format identifier.
1.0
Update IANA Registry with Apple attestation format - The [IANA Registry](https://www.iana.org/assignments/webauthn/webauthn.xhtml) should be updated to include Apple's new attestation format identifier.
process
update iana registry with apple attestation format the should be updated to include apple s new attestation format identifier
1
11,147
13,957,692,968
IssuesEvent
2020-10-24 08:10:50
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
MT - MITA: Harvesting MT
Geoportal Harvesting process MT - Malta
Dear Angelo, Kindly can you perform a harvest for Malta CSW please? as we need to check some changes. Regards, Rene
1.0
MT - MITA: Harvesting MT - Dear Angelo, Kindly can you perform a harvest for Malta CSW please? as we need to check some changes. Regards, Rene
process
mt mita harvesting mt dear angelo kindly can you perform a harvest for malta csw please as we need to check some changes regards rene
1
278,480
30,702,337,984
IssuesEvent
2023-07-27 01:21:43
nidhi7598/linux-3.0.35
https://api.github.com/repos/nidhi7598/linux-3.0.35
closed
CVE-2021-28972 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed
Mend: dependency security vulnerability
## CVE-2021-28972 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In drivers/pci/hotplug/rpadlpar_sysfs.c in the Linux kernel through 5.11.8, the RPA PCI Hotplug driver has a user-tolerable buffer overflow when writing a new device name to the driver from userspace, allowing userspace to write data to the kernel stack frame directly. This occurs because add_slot_store and remove_slot_store mishandle drc_name '\0' termination, aka CID-cc7a0bb058b8. <p>Publish Date: 2021-03-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-28972>CVE-2021-28972</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972</a></p> <p>Release Date: 2021-03-22</p> <p>Fix Resolution: v4.4.263, v4.9.263, v4.14.227, v4.19.183, v5.4.108, v5.10.26, v5.11.9, v5.12-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-28972 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2021-28972 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In drivers/pci/hotplug/rpadlpar_sysfs.c in the Linux kernel through 5.11.8, the RPA PCI Hotplug driver has a user-tolerable buffer overflow when writing a new device name to the driver from userspace, allowing userspace to write data to the kernel stack frame directly. This occurs because add_slot_store and remove_slot_store mishandle drc_name '\0' termination, aka CID-cc7a0bb058b8. <p>Publish Date: 2021-03-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-28972>CVE-2021-28972</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972</a></p> <p>Release Date: 2021-03-22</p> <p>Fix Resolution: v4.4.263, v4.9.263, v4.14.227, v4.19.183, v5.4.108, v5.10.26, v5.11.9, v5.12-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers pci hotplug rpadlpar sysfs c drivers pci hotplug rpadlpar sysfs c drivers pci hotplug rpadlpar sysfs c vulnerability details in drivers pci hotplug rpadlpar sysfs c in the linux kernel through the rpa pci hotplug driver has a user tolerable buffer overflow when writing a new device name to the driver from userspace allowing userspace to write data to the kernel stack frame directly this occurs because add slot store and remove slot store mishandle drc name termination aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
34
2,499,975,319
IssuesEvent
2015-01-08 08:43:15
sysown/proxysql-0.2
https://api.github.com/repos/sysown/proxysql-0.2
opened
Hang on LOAD MYSQL QUERY RULES FROM MEMORY
ADMIN bug QUERY PROCESSOR
``` mysql> select * from mysql_query_rules; Empty set (0.00 sec) mysql> show create table mysql_query_rules\G *************************** 1. row *************************** table: mysql_query_rules Create Table: CREATE TABLE mysql_query_rules (rule_id INTEGER PRIMARY KEY AUTOINCREMENT NOT NULL, active INT CHECK (active IN (0,1)) NOT NULL DEFAULT 0, username VARCHAR, schemaname VARCHAR, flagIN INT NOT NULL DEFAULT 0, match_pattern VARCHAR, negate_match_pattern INT CHECK (negate_match_pattern IN (0,1)) NOT NULL DEFAULT 0, flagOUT INT, replace_pattern VARCHAR, destination_hostgroup INT DEFAULT NULL, cache_ttl INT CHECK(cache_ttl > 0), apply INT CHECK(apply IN (0,1)) NOT NULL DEFAULT 0, FOREIGN KEY (destination_hostgroup) REFERENCES mysql_hostgroups (hostgroup_id)) 1 row in set (0.00 sec) mysql> insert into mysql_query_rules values (1, 1, NULL, NULL, 0, '^SELECT.*', 0, 0, NULL, 0, 10, 1); Query OK, 1 row affected (0.00 sec) mysql> save mysql query rules to disk; Query OK, 0 rows affected (0.01 sec) mysql> load mysql query rules from memory; ```
1.0
Hang on LOAD MYSQL QUERY RULES FROM MEMORY - ``` mysql> select * from mysql_query_rules; Empty set (0.00 sec) mysql> show create table mysql_query_rules\G *************************** 1. row *************************** table: mysql_query_rules Create Table: CREATE TABLE mysql_query_rules (rule_id INTEGER PRIMARY KEY AUTOINCREMENT NOT NULL, active INT CHECK (active IN (0,1)) NOT NULL DEFAULT 0, username VARCHAR, schemaname VARCHAR, flagIN INT NOT NULL DEFAULT 0, match_pattern VARCHAR, negate_match_pattern INT CHECK (negate_match_pattern IN (0,1)) NOT NULL DEFAULT 0, flagOUT INT, replace_pattern VARCHAR, destination_hostgroup INT DEFAULT NULL, cache_ttl INT CHECK(cache_ttl > 0), apply INT CHECK(apply IN (0,1)) NOT NULL DEFAULT 0, FOREIGN KEY (destination_hostgroup) REFERENCES mysql_hostgroups (hostgroup_id)) 1 row in set (0.00 sec) mysql> insert into mysql_query_rules values (1, 1, NULL, NULL, 0, '^SELECT.*', 0, 0, NULL, 0, 10, 1); Query OK, 1 row affected (0.00 sec) mysql> save mysql query rules to disk; Query OK, 0 rows affected (0.01 sec) mysql> load mysql query rules from memory; ```
process
hang on load mysql query rules from memory mysql select from mysql query rules empty set sec mysql show create table mysql query rules g row table mysql query rules create table create table mysql query rules rule id integer primary key autoincrement not null active int check active in not null default username varchar schemaname varchar flagin int not null default match pattern varchar negate match pattern int check negate match pattern in not null default flagout int replace pattern varchar destination hostgroup int default null cache ttl int check cache ttl apply int check apply in not null default foreign key destination hostgroup references mysql hostgroups hostgroup id row in set sec mysql insert into mysql query rules values null null select null query ok row affected sec mysql save mysql query rules to disk query ok rows affected sec mysql load mysql query rules from memory
1
190,350
6,817,708,068
IssuesEvent
2017-11-07 00:50:58
FRCteam4909/TGA-Server-2018
https://api.github.com/repos/FRCteam4909/TGA-Server-2018
opened
Analysis Portal & BT Worker Integration (Server->Tablet)
Priority: High Status: Assigned Type: Enhancement
The analysis portal should be based on Node, and Mongo as per #2. This should connect with the BT worker via something similar to web hooks, both ways. Tablet->Server has since been implemented in #4.
1.0
Analysis Portal & BT Worker Integration (Server->Tablet) - The analysis portal should be based on Node, and Mongo as per #2. This should connect with the BT worker via something similar to web hooks, both ways. Tablet->Server has since been implemented in #4.
non_process
analysis portal bt worker integration server tablet the analysis portal should be based on node and mongo as per this should connect with the bt worker via something similar to web hooks both ways tablet server has since been implemented in
0
591,007
17,792,932,822
IssuesEvent
2021-08-31 18:24:00
status-im/status-desktop
https://api.github.com/repos/status-im/status-desktop
closed
no identicon/profile picture shown and wrong timestamp format in search results
bug ui priority F2: important
Search results should display the identicon or profile picture if available. Also timestamp is shown as a number instead of correct date format. ![image](https://user-images.githubusercontent.com/1106587/130239409-ff9f6f46-e08c-4597-9015-c6178c1cf2c9.png)
1.0
no identicon/profile picture shown and wrong timestamp format in search results - Search results should display the identicon or profile picture if available. Also timestamp is shown as a number instead of correct date format. ![image](https://user-images.githubusercontent.com/1106587/130239409-ff9f6f46-e08c-4597-9015-c6178c1cf2c9.png)
non_process
no identicon profile picture shown and wrong timestamp format in search results search results should display the identicon or profile picture if available also timestamp is shown as a number instead of correct date format
0
20,225
3,799,106,924
IssuesEvent
2016-03-23 14:57:01
ensime/scala-debugger
https://api.github.com/repos/ensime/scala-debugger
closed
Create Drone cache image for testing
enhancement for test Needs Review
Currently, the Drone setup downloads the sbt plugins (including Coursier) as well as all dependencies on each run. It would be a very good idea to produce a custom Docker image that contains these dependencies to both speed up run time and avoid hammering Maven Central (and other repositories). I'd love for the auto-commit caching to work (it looks like Drone is doing something to do with committing a new image), but for now this seems to be the best approach. This is what you were talking about earlier, right, @fommil?
1.0
Create Drone cache image for testing - Currently, the Drone setup downloads the sbt plugins (including Coursier) as well as all dependencies on each run. It would be a very good idea to produce a custom Docker image that contains these dependencies to both speed up run time and avoid hammering Maven Central (and other repositories). I'd love for the auto-commit caching to work (it looks like Drone is doing something to do with committing a new image), but for now this seems to be the best approach. This is what you were talking about earlier, right, @fommil?
non_process
create drone cache image for testing currently the drone setup downloads the sbt plugins including coursier as well as all dependencies on each run it would be a very good idea to produce a custom docker image that contains these dependencies to both speed up run time and avoid hammering maven central and other repositories i d love for the auto commit caching to work it looks like drone is doing something to do with committing a new image but for now this seems to be the best approach this is what you were talking about earlier right fommil
0
13,590
16,162,951,516
IssuesEvent
2021-05-01 01:27:17
tdwg/chrono
https://api.github.com/repos/tdwg/chrono
closed
Handling of IRI values
Process - prepare for Executive review Question - answered
Is there a mechanism to supply IRI values for chrono:chronometricAgeDeterminedBy, either with an id term chrono:chronometricAgeDeterminedByID or an IRI namespace chronoiri:chronometricAgeDeterminedBy? Such a mechanism would be needed to allow an RDF representation to express a list of chronometric age determiners as repeated id terms, each pointing to a determiner agent by ORCID or other agent identifier. _Originally posted by @chicoreus in https://github.com/tdwg/chrono/issues/15#issuecomment-732198397_
1.0
Handling of IRI values - Is there a mechanism to supply IRI values for chrono:chronometricAgeDeterminedBy, either with an id term chrono:chronometricAgeDeterminedByID or an IRI namespace chronoiri:chronometricAgeDeterminedBy? Such a mechanism would be needed to allow an RDF representation to express a list of chronometric age determiners as repeated id terms, each pointing to a determiner agent by ORCID or other agent identifier. _Originally posted by @chicoreus in https://github.com/tdwg/chrono/issues/15#issuecomment-732198397_
process
handling of iri values is there a mechanism to supply iri values for chrono chronometricagedeterminedby either with an id term chrono chronometricagedeterminedbyid or an iri namespace chronoiri chronometricagedeterminedby such a mechanism would be needed to allow an rdf representation to express a list of chronometric age determiners as repeated id terms each pointing to a determiner agent by orcid or other agent identifier originally posted by chicoreus in
1
21,656
30,105,856,141
IssuesEvent
2023-06-30 01:07:53
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
hpcflow-new2 0.2.0a55 has 1 GuardDog issues
guarddog silent-process-execution
https://pypi.org/project/hpcflow-new2 https://inspector.pypi.io/project/hpcflow-new2 ```{ "dependency": "hpcflow-new2", "version": "0.2.0a55", "result": { "issues": 1, "errors": {}, "results": { "silent-process-execution": [ { "location": "hpcflow_new2-0.2.0a55/hpcflow/sdk/helper/helper.py:118", "code": " proc = subprocess.Popen(\n args=args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n **kwargs,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpbko46hx_/hpcflow-new2" } }```
1.0
hpcflow-new2 0.2.0a55 has 1 GuardDog issues - https://pypi.org/project/hpcflow-new2 https://inspector.pypi.io/project/hpcflow-new2 ```{ "dependency": "hpcflow-new2", "version": "0.2.0a55", "result": { "issues": 1, "errors": {}, "results": { "silent-process-execution": [ { "location": "hpcflow_new2-0.2.0a55/hpcflow/sdk/helper/helper.py:118", "code": " proc = subprocess.Popen(\n args=args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n **kwargs,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpbko46hx_/hpcflow-new2" } }```
process
hpcflow has guarddog issues dependency hpcflow version result issues errors results silent process execution location hpcflow hpcflow sdk helper helper py code proc subprocess popen n args args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n kwargs n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp hpcflow
1
790,943
27,843,833,896
IssuesEvent
2023-03-20 14:22:56
GSM-MSG/GCMS-FrontEnd-V2
https://api.github.com/repos/GSM-MSG/GCMS-FrontEnd-V2
closed
동아리 디테일에서 설명이나 제목이 space나 \n을 안 읽어요
2️⃣ Priority: Medium ♻️ Refactor 🐞 Bug
### Describe <img width="1392" alt="스크린샷 2023-03-20 오전 1 15 55" src="https://user-images.githubusercontent.com/81547954/226189371-f2c619ef-847b-4ba7-a2aa-fe3f122fe558.png"> 보세요! ### Additional _No response_
1.0
동아리 디테일에서 설명이나 제목이 space나 \n을 안 읽어요 - ### Describe <img width="1392" alt="스크린샷 2023-03-20 오전 1 15 55" src="https://user-images.githubusercontent.com/81547954/226189371-f2c619ef-847b-4ba7-a2aa-fe3f122fe558.png"> 보세요! ### Additional _No response_
non_process
동아리 디테일에서 설명이나 제목이 space나 n을 안 읽어요 describe img width alt 스크린샷 오전 src 보세요 additional no response
0
14,304
17,290,977,235
IssuesEvent
2021-07-24 18:46:56
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
@parcel/transformer-sass: Fails to resolve path to node module
:bug: Bug CSS Preprocessing ✨ Parcel 2
<!--- Thanks for filing an issue 😄 ! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> # 🐛 bug report In the newly released parcel 2.0.0-beta.2 (and nightly builds before that) transformer-sass can’t resolve paths to installed node modules. The error looks similar to https://github.com/parcel-bundler/parcel/issues/5813 but happens on build and serve. ## 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> package.json ```js "dependencies": { "autoprefixer": "~10.2.5", "core-js": "~3.9.1", "foundation-sites": "6.6.3", "husky": "~5.1.3", "jquery": "~3.6.0", "motion-ui": "~2.0.3", "parcel": "2.0.0-beta.2", "postcss": "~8.2.7", "sass": "~1.32.8", }, // automatically added by parcel: "devDependencies": { "@babel/core": "^7.13.10", "@parcel/transformer-sass": "2.0.0-beta.2" } ``` .scssrc ```js { "includePaths": ["node_modules"] } ``` In my scss files, I'm importing scss files from foundation-sites, e.g. ```scss @import "foundation-sites/scss/global"; ``` ## 🤔 Expected Behavior I expect the scss file to be included (used to still work with parcel 2.0.0-beta.1). ## 😯 Current Behavior <!--- Tell us what happens instead of the expected behavior --> `yarn parcel serve` and `yarn parcel build` throw an error. No css file is build. <!--- If you are seeing an error, please include the full error message and stack trace --> ``` Build failed. @parcel/transformer-sass: Got unexpected null ╷ 19 │ @import "foundation-sites/scss/global"; │ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ╵ scss/main.scss 19:9 root stylesheet Error: Got unexpected null ╷ 19 │ @import "foundation-sites/scss/global"; │ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ╵ scss/main.scss 19:9 root stylesheet at Object._newRenderError (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13537:19) at Object._wrapException (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13374:16) at _render_closure1.call$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:80373:21) at _RootZone.runBinary$3$3 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:27269:18) at _FutureListener.handleError$1 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25797:19) at _Future__propagateToListeners_handleError.call$0 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:26094:49) at Object._Future__propagateToListeners (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4543:77) at _Future._completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25927:9) at _AsyncAwaitCompleter.completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25270:12) at Object._asyncRethrow (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4292:17) ``` On previous tries the error message was "Failed to resolve" with an identical stack trace. ## 💁 Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug --> Maybe scss files need to be imported differently now? ## 🔦 Context <!--- How has this issue affected you? What are you trying to accomplish? --> The issue makes any recent version of parcel 2 unusable for us. We need to be able to import scss files from other node modules. <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## 💻 Code Sample <!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue --> ## 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 2.0.0-beta.2 | Node | 12.19.0 | npm/Yarn | yarn 1.22.5 | Operating System | Ubuntu 20.10 <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
1.0
@parcel/transformer-sass: Fails to resolve path to node module - <!--- Thanks for filing an issue 😄 ! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> # 🐛 bug report In the newly released parcel 2.0.0-beta.2 (and nightly builds before that) transformer-sass can’t resolve paths to installed node modules. The error looks similar to https://github.com/parcel-bundler/parcel/issues/5813 but happens on build and serve. ## 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> package.json ```js "dependencies": { "autoprefixer": "~10.2.5", "core-js": "~3.9.1", "foundation-sites": "6.6.3", "husky": "~5.1.3", "jquery": "~3.6.0", "motion-ui": "~2.0.3", "parcel": "2.0.0-beta.2", "postcss": "~8.2.7", "sass": "~1.32.8", }, // automatically added by parcel: "devDependencies": { "@babel/core": "^7.13.10", "@parcel/transformer-sass": "2.0.0-beta.2" } ``` .scssrc ```js { "includePaths": ["node_modules"] } ``` In my scss files, I'm importing scss files from foundation-sites, e.g. ```scss @import "foundation-sites/scss/global"; ``` ## 🤔 Expected Behavior I expect the scss file to be included (used to still work with parcel 2.0.0-beta.1). ## 😯 Current Behavior <!--- Tell us what happens instead of the expected behavior --> `yarn parcel serve` and `yarn parcel build` throw an error. No css file is build. <!--- If you are seeing an error, please include the full error message and stack trace --> ``` Build failed. @parcel/transformer-sass: Got unexpected null ╷ 19 │ @import "foundation-sites/scss/global"; │ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ╵ scss/main.scss 19:9 root stylesheet Error: Got unexpected null ╷ 19 │ @import "foundation-sites/scss/global"; │ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ╵ scss/main.scss 19:9 root stylesheet at Object._newRenderError (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13537:19) at Object._wrapException (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13374:16) at _render_closure1.call$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:80373:21) at _RootZone.runBinary$3$3 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:27269:18) at _FutureListener.handleError$1 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25797:19) at _Future__propagateToListeners_handleError.call$0 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:26094:49) at Object._Future__propagateToListeners (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4543:77) at _Future._completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25927:9) at _AsyncAwaitCompleter.completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25270:12) at Object._asyncRethrow (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4292:17) ``` On previous tries the error message was "Failed to resolve" with an identical stack trace. ## 💁 Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug --> Maybe scss files need to be imported differently now? ## 🔦 Context <!--- How has this issue affected you? What are you trying to accomplish? --> The issue makes any recent version of parcel 2 unusable for us. We need to be able to import scss files from other node modules. <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## 💻 Code Sample <!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue --> ## 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 2.0.0-beta.2 | Node | 12.19.0 | npm/Yarn | yarn 1.22.5 | Operating System | Ubuntu 20.10 <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
process
parcel transformer sass fails to resolve path to node module thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before 🐛 bug report in the newly released parcel beta and nightly builds before that transformer sass can’t resolve paths to installed node modules the error looks similar to but happens on build and serve 🎛 configuration babelrc package json cli command package json js dependencies autoprefixer core js foundation sites husky jquery motion ui parcel beta postcss sass automatically added by parcel devdependencies babel core parcel transformer sass beta scssrc js includepaths in my scss files i m importing scss files from foundation sites e g scss import foundation sites scss global 🤔 expected behavior i expect the scss file to be included used to still work with parcel beta 😯 current behavior yarn parcel serve and yarn parcel build throw an error no css file is build build failed parcel transformer sass got unexpected null ╷ │ import foundation sites scss global │ ╵ scss main scss root stylesheet error got unexpected null ╷ │ import foundation sites scss global │ ╵ scss main scss root stylesheet at object newrendererror var code themes mo foundation base node modules sass sass dart js at object wrapexception var code themes mo foundation base node modules sass sass dart js at render call var code themes mo foundation base node modules sass sass dart js at rootzone runbinary var code themes mo foundation base node modules sass sass dart js at futurelistener handleerror var code themes mo foundation base node modules sass sass dart js at future propagatetolisteners handleerror call var code themes mo foundation base node modules sass sass dart js at object future propagatetolisteners var code themes mo foundation base node modules sass sass dart js at future completeerror var code themes mo foundation base node modules sass sass dart js at asyncawaitcompleter completeerror var code themes mo foundation base node modules sass sass dart js at object asyncrethrow var code themes mo foundation base node modules sass sass dart js on previous tries the error message was failed to resolve with an identical stack trace 💁 possible solution maybe scss files need to be imported differently now 🔦 context the issue makes any recent version of parcel unusable for us we need to be able to import scss files from other node modules 💻 code sample 🌍 your environment software version s parcel beta node npm yarn yarn operating system ubuntu love parcel please consider supporting our collective 👉
1
55,860
3,075,045,331
IssuesEvent
2015-08-20 11:16:02
cdnjs/cdnjs
https://api.github.com/repos/cdnjs/cdnjs
closed
[bug]npm auto-update can not pick every versions properly at once
Bug - High Priority Help wanted
Need to run the auto-update process for many times to pick all the versions > error downloading `123` of `xxx` it didnt exist: [object Object]null
1.0
[bug]npm auto-update can not pick every versions properly at once - Need to run the auto-update process for many times to pick all the versions > error downloading `123` of `xxx` it didnt exist: [object Object]null
non_process
npm auto update can not pick every versions properly at once need to run the auto update process for many times to pick all the versions error downloading of xxx it didnt exist null
0
7,757
10,878,344,777
IssuesEvent
2019-11-16 17:00:52
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Refactor workflow in App-Frontend
area/process area/receipt breaking-change kind/user-story solution/app-frontend team/tamagotchi
## Description Based on the analysis in #13678, refactor Workflowstep / Formfiller / Receipt ## BREAKING CHANGE Note that this issue introduces breaking changes for existing applications. How to fix is documented [on altinn docs.](https://docs.altinn.studio/known-issues/breaking-changes/#breaking-change-data-element-validation-returns-an-internal-server-error-to-app-frontend) ## Desired process flow ![image](https://user-images.githubusercontent.com/10416981/68115680-3c179100-fef9-11e9-869b-673e2fe3e3a0.png) ## Considerations Input (beyond tasks) on how the user story should be solved can be put here. ## Acceptance criteria - App frontend are updated to use new process api - The 'Send inn' step is removed - All other functionality works as before ## Out of scope - Generic process handling is scoped out ## Tasks - [x] Verify that this issue meets DoP (remove unused text, add missing text/parameters/labels, verify tasks). ### Specification tasks - [x] Test design / decide test need ### Development tasks - [x] Refactor the "workflow resource" to `shared/resources/process` - [x] Move src\react-apps\applications\runtime\src\features\form\workflow - [x] Refactor src\react-apps\applications\runtime\src\containers\WorkflowStep.tsx - [x] Rename "set current state" action to "write current workflow state to store" or similar... *needs to be discussed* - [x] Actions,Sagas,Reducers from src\react-apps\applications\runtime\src\features\form\workflow\ needs to be moved to new folder with new structure. - [x] Create `getProcessStateSaga` that gets current process state. Place under `shared/resources/process/getProcessState` - [x] Update how we handle "Update" and "Complete" actions to the controller @lorang92 - [x] Update should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/data/{dataGuid}` - [x] Complete should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/process/completeProcess` - [x] Create shared ProcessStep enum - [x] Rename Workflow containers/components to Process ### QA - [x] Documentation (if relevant) - [x] Manual test (if needed) - [x] Automated test (if needed)
1.0
Refactor workflow in App-Frontend - ## Description Based on the analysis in #13678, refactor Workflowstep / Formfiller / Receipt ## BREAKING CHANGE Note that this issue introduces breaking changes for existing applications. How to fix is documented [on altinn docs.](https://docs.altinn.studio/known-issues/breaking-changes/#breaking-change-data-element-validation-returns-an-internal-server-error-to-app-frontend) ## Desired process flow ![image](https://user-images.githubusercontent.com/10416981/68115680-3c179100-fef9-11e9-869b-673e2fe3e3a0.png) ## Considerations Input (beyond tasks) on how the user story should be solved can be put here. ## Acceptance criteria - App frontend are updated to use new process api - The 'Send inn' step is removed - All other functionality works as before ## Out of scope - Generic process handling is scoped out ## Tasks - [x] Verify that this issue meets DoP (remove unused text, add missing text/parameters/labels, verify tasks). ### Specification tasks - [x] Test design / decide test need ### Development tasks - [x] Refactor the "workflow resource" to `shared/resources/process` - [x] Move src\react-apps\applications\runtime\src\features\form\workflow - [x] Refactor src\react-apps\applications\runtime\src\containers\WorkflowStep.tsx - [x] Rename "set current state" action to "write current workflow state to store" or similar... *needs to be discussed* - [x] Actions,Sagas,Reducers from src\react-apps\applications\runtime\src\features\form\workflow\ needs to be moved to new folder with new structure. - [x] Create `getProcessStateSaga` that gets current process state. Place under `shared/resources/process/getProcessState` - [x] Update how we handle "Update" and "Complete" actions to the controller @lorang92 - [x] Update should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/data/{dataGuid}` - [x] Complete should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/process/completeProcess` - [x] Create shared ProcessStep enum - [x] Rename Workflow containers/components to Process ### QA - [x] Documentation (if relevant) - [x] Manual test (if needed) - [x] Automated test (if needed)
process
refactor workflow in app frontend description based on the analysis in refactor workflowstep formfiller receipt breaking change note that this issue introduces breaking changes for existing applications how to fix is documented desired process flow considerations input beyond tasks on how the user story should be solved can be put here acceptance criteria app frontend are updated to use new process api the send inn step is removed all other functionality works as before out of scope generic process handling is scoped out tasks verify that this issue meets dop remove unused text add missing text parameters labels verify tasks specification tasks test design decide test need development tasks refactor the workflow resource to shared resources process move src react apps applications runtime src features form workflow refactor src react apps applications runtime src containers workflowstep tsx rename set current state action to write current workflow state to store or similar needs to be discussed actions sagas reducers from src react apps applications runtime src features form workflow needs to be moved to new folder with new structure create getprocessstatesaga that gets current process state place under shared resources process getprocessstate update how we handle update and complete actions to the controller update should call org app instances instanceownerid instanceguid data dataguid complete should call org app instances instanceownerid instanceguid process completeprocess create shared processstep enum rename workflow containers components to process qa documentation if relevant manual test if needed automated test if needed
1
822,994
30,923,281,124
IssuesEvent
2023-08-06 06:56:38
unitystation/unitystation
https://api.github.com/repos/unitystation/unitystation
closed
The text size in the player list is massive
Type: Feature Priority: Low Good First Issue
Thanks to the text size, the player list can only fit something like five people at the moment. We might wanna scale that down just a tad. https://i.imgur.com/un9DBVd.png
1.0
The text size in the player list is massive - Thanks to the text size, the player list can only fit something like five people at the moment. We might wanna scale that down just a tad. https://i.imgur.com/un9DBVd.png
non_process
the text size in the player list is massive thanks to the text size the player list can only fit something like five people at the moment we might wanna scale that down just a tad
0
399,358
11,747,581,358
IssuesEvent
2020-03-12 13:52:37
AbsaOSS/hyperdrive-trigger
https://api.github.com/repos/AbsaOSS/hyperdrive-trigger
opened
Tables - use identity instead serial in table implementations
backend bug priority: medium
Instead of serial, you should probably use identity, since identity is an SQL-Standard whereas serial is postgres specific. E.g. "id" BIGINT GENERATED BY DEFAULT AS IDENTITY NOT NULL PRIMARY KEY
1.0
Tables - use identity instead serial in table implementations - Instead of serial, you should probably use identity, since identity is an SQL-Standard whereas serial is postgres specific. E.g. "id" BIGINT GENERATED BY DEFAULT AS IDENTITY NOT NULL PRIMARY KEY
non_process
tables use identity instead serial in table implementations instead of serial you should probably use identity since identity is an sql standard whereas serial is postgres specific e g id bigint generated by default as identity not null primary key
0
21,549
29,865,410,710
IssuesEvent
2023-06-20 03:04:17
cncf/tag-security
https://api.github.com/repos/cncf/tag-security
closed
[Sec Assess WG] Mapping Security Assessments to TOC Process
help wanted assessment-process suggestion inactive
This issue was created from results of the Security Assessment Improvement Working Group (https://github.com/cncf/sig-security/issues/167#issuecomment-714514142). # Mapping Security Assessments to TOC Process ## Premise - The security assessment aims to tie into the CNCF project process, but it is not clear how - There is no current agreed upon requirement for the process for what is asked of a SIG ## Ideas - better document the Due Diligence process and document - Need more detail on when the SIG recommends for the next CNCF phase - not tie it too directly to TOC process - CNCF needs to explicitly define the requirement for projects at each level to go through assessment - Map security assessment process to process of CNCF - Are assessments necessary/mandated by the TOC to move from one stage to another ? We need to make projects realize it's in their interest to invest time in going through the assessment process. ## Action Items - [ ] Create internal mapping of TOC mapping for SIG-Security (@lumjjb, @itaysk) - [ ] Propose a hard requirement for TOC process (looking at frequency/data of incubating/graduating projects) ## Logistics - [ ] Contributors (For multiple contributors, 1 lead to coordinate) - [ ] SIG-Representative
1.0
[Sec Assess WG] Mapping Security Assessments to TOC Process - This issue was created from results of the Security Assessment Improvement Working Group (https://github.com/cncf/sig-security/issues/167#issuecomment-714514142). # Mapping Security Assessments to TOC Process ## Premise - The security assessment aims to tie into the CNCF project process, but it is not clear how - There is no current agreed upon requirement for the process for what is asked of a SIG ## Ideas - better document the Due Diligence process and document - Need more detail on when the SIG recommends for the next CNCF phase - not tie it too directly to TOC process - CNCF needs to explicitly define the requirement for projects at each level to go through assessment - Map security assessment process to process of CNCF - Are assessments necessary/mandated by the TOC to move from one stage to another ? We need to make projects realize it's in their interest to invest time in going through the assessment process. ## Action Items - [ ] Create internal mapping of TOC mapping for SIG-Security (@lumjjb, @itaysk) - [ ] Propose a hard requirement for TOC process (looking at frequency/data of incubating/graduating projects) ## Logistics - [ ] Contributors (For multiple contributors, 1 lead to coordinate) - [ ] SIG-Representative
process
mapping security assessments to toc process this issue was created from results of the security assessment improvement working group mapping security assessments to toc process premise the security assessment aims to tie into the cncf project process but it is not clear how there is no current agreed upon requirement for the process for what is asked of a sig ideas better document the due diligence process and document need more detail on when the sig recommends for the next cncf phase not tie it too directly to toc process cncf needs to explicitly define the requirement for projects at each level to go through assessment map security assessment process to process of cncf are assessments necessary mandated by the toc to move from one stage to another we need to make projects realize it s in their interest to invest time in going through the assessment process action items create internal mapping of toc mapping for sig security lumjjb itaysk propose a hard requirement for toc process looking at frequency data of incubating graduating projects logistics contributors for multiple contributors lead to coordinate sig representative
1
15,576
19,703,508,290
IssuesEvent
2022-01-12 19:08:19
googleapis/nodejs-filestore
https://api.github.com/repos/googleapis/nodejs-filestore
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'filestore' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'filestore' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname filestore invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
14,714
17,926,562,100
IssuesEvent
2021-09-10 00:51:39
Leviatan-Analytics/LA-data-processing
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
closed
Implement text recognition and ward detection scripts parallelization [3]
Data Processing Week 2 Sprint 4
Use the knowledge of the previous research and implement the parallelization of the scripts.
1.0
Implement text recognition and ward detection scripts parallelization [3] - Use the knowledge of the previous research and implement the parallelization of the scripts.
process
implement text recognition and ward detection scripts parallelization use the knowledge of the previous research and implement the parallelization of the scripts
1
55,837
11,470,803,925
IssuesEvent
2020-02-09 06:27:18
eclipse-theia/theia
https://api.github.com/repos/eclipse-theia/theia
opened
plugin tree rows are bogusly rendered
tree ui/ux vscode
Some recent PR, broke plugin tree rendering, rows are not rendered inline anymore <img width="1904" alt="Screenshot 2020-02-09 at 07 25 26" src="https://user-images.githubusercontent.com/3082655/74097556-7fdbf100-4b0d-11ea-9c27-f7b8c934fb47.png"> @vince-fugnitto Do you remember which PR? Could we revert it? It is quite bad now.
1.0
plugin tree rows are bogusly rendered - Some recent PR, broke plugin tree rendering, rows are not rendered inline anymore <img width="1904" alt="Screenshot 2020-02-09 at 07 25 26" src="https://user-images.githubusercontent.com/3082655/74097556-7fdbf100-4b0d-11ea-9c27-f7b8c934fb47.png"> @vince-fugnitto Do you remember which PR? Could we revert it? It is quite bad now.
non_process
plugin tree rows are bogusly rendered some recent pr broke plugin tree rendering rows are not rendered inline anymore img width alt screenshot at src vince fugnitto do you remember which pr could we revert it it is quite bad now
0
14,121
10,616,442,453
IssuesEvent
2019-10-12 11:46:24
python-pillow/Pillow
https://api.github.com/repos/python-pillow/Pillow
closed
Automated security fixes with Dependabot?
Infrastructure
GitHub will be sending automated PRs with security fixes for accounts with security alerts enabled. Pillow does not have any Python dependencies for production code, only testing, but shall we enabled the setting to see how it goes? It's at https://github.com/python-pillow/Pillow/network/alerts (I don't have access). More info: * https://github.blog/2019-05-23-introducing-new-ways-to-keep-your-code-secure/#automated-security-fixes-with-dependabot * https://help.github.com/en/articles/configuring-automated-security-fixes
1.0
Automated security fixes with Dependabot? - GitHub will be sending automated PRs with security fixes for accounts with security alerts enabled. Pillow does not have any Python dependencies for production code, only testing, but shall we enabled the setting to see how it goes? It's at https://github.com/python-pillow/Pillow/network/alerts (I don't have access). More info: * https://github.blog/2019-05-23-introducing-new-ways-to-keep-your-code-secure/#automated-security-fixes-with-dependabot * https://help.github.com/en/articles/configuring-automated-security-fixes
non_process
automated security fixes with dependabot github will be sending automated prs with security fixes for accounts with security alerts enabled pillow does not have any python dependencies for production code only testing but shall we enabled the setting to see how it goes it s at i don t have access more info
0
5,689
8,560,128,212
IssuesEvent
2018-11-08 23:44:45
knative/serving
https://api.github.com/repos/knative/serving
closed
Upgrade testing
area/API area/test-and-release kind/feature kind/process
<!-- /area API /area test-and-release /kind feature /kind process /assign @jonjohnsonjr --> We should have testing that verifies that upgrading from our last release to HEAD works.
1.0
Upgrade testing - <!-- /area API /area test-and-release /kind feature /kind process /assign @jonjohnsonjr --> We should have testing that verifies that upgrading from our last release to HEAD works.
process
upgrade testing area api area test and release kind feature kind process assign jonjohnsonjr we should have testing that verifies that upgrading from our last release to head works
1
16,713
21,871,329,092
IssuesEvent
2022-05-19 05:45:10
keras-team/keras-cv
https://api.github.com/repos/keras-team/keras-cv
closed
`MaybeApply` high order abstraction
contribution-welcome preprocessing
We should consider an implementation for a `MaybeApply` layer: ``` maybe_solarize = MaybeApply(Solarization(), rate=0.5) # maybe_solarize solarizes half of the images passed to it ``` https://github.com/chjort/chambers/blob/b796f507b7353fb141b5c8b026b811a023d303e6/chambers/augmentations/image_augmentations.py#L514 Notes: - we can rely on this in RandomAugmentationPipeline - this must be element-wise, not batch wise!
1.0
`MaybeApply` high order abstraction - We should consider an implementation for a `MaybeApply` layer: ``` maybe_solarize = MaybeApply(Solarization(), rate=0.5) # maybe_solarize solarizes half of the images passed to it ``` https://github.com/chjort/chambers/blob/b796f507b7353fb141b5c8b026b811a023d303e6/chambers/augmentations/image_augmentations.py#L514 Notes: - we can rely on this in RandomAugmentationPipeline - this must be element-wise, not batch wise!
process
maybeapply high order abstraction we should consider an implementation for a maybeapply layer maybe solarize maybeapply solarization rate maybe solarize solarizes half of the images passed to it notes we can rely on this in randomaugmentationpipeline this must be element wise not batch wise
1
194,963
22,281,615,794
IssuesEvent
2022-06-11 01:14:31
pazhanivel07/linux-4.19.72
https://api.github.com/repos/pazhanivel07/linux-4.19.72
reopened
CVE-2021-20177 (Medium) detected in linux-yoctov5.4.51, linux-yoctov5.4.51
security vulnerability
## CVE-2021-20177 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linux-yoctov5.4.51</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Linux kernel's implementation of string matching within a packet. A privileged user (with root or CAP_NET_ADMIN) when inserting iptables rules could insert a rule which can panic the system. Kernel before kernel 5.5-rc1 is affected. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20177>CVE-2021-20177</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-20177">https://www.linuxkernelcves.com/cves/CVE-2021-20177</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: v5.5-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-20177 (Medium) detected in linux-yoctov5.4.51, linux-yoctov5.4.51 - ## CVE-2021-20177 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linux-yoctov5.4.51</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Linux kernel's implementation of string matching within a packet. A privileged user (with root or CAP_NET_ADMIN) when inserting iptables rules could insert a rule which can panic the system. Kernel before kernel 5.5-rc1 is affected. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20177>CVE-2021-20177</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-20177">https://www.linuxkernelcves.com/cves/CVE-2021-20177</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: v5.5-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux linux cve medium severity vulnerability vulnerable libraries linux linux vulnerability details a flaw was found in the linux kernel s implementation of string matching within a packet a privileged user with root or cap net admin when inserting iptables rules could insert a rule which can panic the system kernel before kernel is affected publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
3,451
6,542,244,402
IssuesEvent
2017-09-02 02:54:51
gaocegege/Processing.R
https://api.github.com/repos/gaocegege/Processing.R
closed
Replace ant with maven or gradle
community/processing priority/p3 size/no-idea status/to-be-claimed type/feasibility-research
Now Processing.R uses maven-ant-tasks to install maven dependencies in ant. If we could use maven instead, it may be better.
1.0
Replace ant with maven or gradle - Now Processing.R uses maven-ant-tasks to install maven dependencies in ant. If we could use maven instead, it may be better.
process
replace ant with maven or gradle now processing r uses maven ant tasks to install maven dependencies in ant if we could use maven instead it may be better
1
19,311
25,466,751,008
IssuesEvent
2022-11-25 05:43:13
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IDP] [PM] UI Issue when user enters wrong password in the auth server
Bug P1 Process: Fixed Process: Tested QA Process: Tested dev Auth server
UI Issue when user enters wrong password in the 'AUTH' server **AR:** ![UI PM](https://user-images.githubusercontent.com/86007179/199964208-754f65de-5630-49ff-9441-2e1f9507b334.png) **ER:** ![image](https://user-images.githubusercontent.com/86007179/199963807-b20201fb-fcb0-4a04-96fd-e1eb5985a8a3.png)
3.0
[IDP] [PM] UI Issue when user enters wrong password in the auth server - UI Issue when user enters wrong password in the 'AUTH' server **AR:** ![UI PM](https://user-images.githubusercontent.com/86007179/199964208-754f65de-5630-49ff-9441-2e1f9507b334.png) **ER:** ![image](https://user-images.githubusercontent.com/86007179/199963807-b20201fb-fcb0-4a04-96fd-e1eb5985a8a3.png)
process
ui issue when user enters wrong password in the auth server ui issue when user enters wrong password in the auth server ar er
1
367,354
25,734,280,477
IssuesEvent
2022-12-07 23:01:51
cncf/cnf-wg
https://api.github.com/repos/cncf/cnf-wg
opened
Suggestion: Update categories names to be the same types of words
documentation enhancement
Some are properties, some are areas, etc. Suggestion: use words that are of one type for each category name. Examples: - [Observability and Diagnostics](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#90-observability-and-diagnostics) - [Security](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#70-security)
1.0
Suggestion: Update categories names to be the same types of words - Some are properties, some are areas, etc. Suggestion: use words that are of one type for each category name. Examples: - [Observability and Diagnostics](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#90-observability-and-diagnostics) - [Security](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#70-security)
non_process
suggestion update categories names to be the same types of words some are properties some are areas etc suggestion use words that are of one type for each category name examples
0
2,436
5,216,170,327
IssuesEvent
2017-01-26 09:22:15
QCoDeS/Qcodes
https://api.github.com/repos/QCoDeS/Qcodes
closed
PyQtgraph live plot does not update
bug mulitprocessing
Issue made to get rid of the hack in PR #250 . ### Steps to reproduce 1. Run Tutorial notebook from #250 ``` python data = qc.Loop(c0.sweep(-20,20,0.1), delay=0.003).each(meter.amplitude).run(name='testsweep') plotQ = qc.QtPlot() plotQ.add(data.meter_amplitude) ``` ### Expected behaviour Expect to see a live updating plot of the data ### Actual behaviour Plots only what is in "data" the moment the plot function is called. ### Workaround ``` python while data.sync(): plotQ.update() plotQ.update() ``` The following code will keep updating the plot (in the main thread) while the data taking is not finished yet. ### System **operating system** OS X El Capitan 10.11.5 (15F34) **qcodes branch** "examples" (only contains notebook so should be close to master) **qcodes commit** Commit: 007e5f95ac5
1.0
PyQtgraph live plot does not update - Issue made to get rid of the hack in PR #250 . ### Steps to reproduce 1. Run Tutorial notebook from #250 ``` python data = qc.Loop(c0.sweep(-20,20,0.1), delay=0.003).each(meter.amplitude).run(name='testsweep') plotQ = qc.QtPlot() plotQ.add(data.meter_amplitude) ``` ### Expected behaviour Expect to see a live updating plot of the data ### Actual behaviour Plots only what is in "data" the moment the plot function is called. ### Workaround ``` python while data.sync(): plotQ.update() plotQ.update() ``` The following code will keep updating the plot (in the main thread) while the data taking is not finished yet. ### System **operating system** OS X El Capitan 10.11.5 (15F34) **qcodes branch** "examples" (only contains notebook so should be close to master) **qcodes commit** Commit: 007e5f95ac5
process
pyqtgraph live plot does not update issue made to get rid of the hack in pr steps to reproduce run tutorial notebook from python data qc loop sweep delay each meter amplitude run name testsweep plotq qc qtplot plotq add data meter amplitude expected behaviour expect to see a live updating plot of the data actual behaviour plots only what is in data the moment the plot function is called workaround python while data sync plotq update plotq update the following code will keep updating the plot in the main thread while the data taking is not finished yet system operating system os x el capitan qcodes branch examples only contains notebook so should be close to master qcodes commit commit
1
46,678
7,275,022,638
IssuesEvent
2018-02-21 12:06:30
rucio/rucio
https://api.github.com/repos/rucio/rucio
closed
Broken readthedocs compilation with No such file or directory
Documentation
Motivation ---------- ``` File "/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/envs/latest/local/lib/python2.7/site-packages/rucio-1.14.10-py2.7.egg/rucio/common/doc/argparse/ext.py", line 415, in run f = open(os.path.abspath(self.options['filename'])) IOError: [Errno 2] No such file or directory: u'/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/checkouts/latest/doc/source/bin/rucio' ```
1.0
Broken readthedocs compilation with No such file or directory - Motivation ---------- ``` File "/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/envs/latest/local/lib/python2.7/site-packages/rucio-1.14.10-py2.7.egg/rucio/common/doc/argparse/ext.py", line 415, in run f = open(os.path.abspath(self.options['filename'])) IOError: [Errno 2] No such file or directory: u'/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/checkouts/latest/doc/source/bin/rucio' ```
non_process
broken readthedocs compilation with no such file or directory motivation file home docs checkouts readthedocs org user builds rucio blueprint envs latest local lib site packages rucio egg rucio common doc argparse ext py line in run f open os path abspath self options ioerror no such file or directory u home docs checkouts readthedocs org user builds rucio blueprint checkouts latest doc source bin rucio
0
324,214
9,886,276,587
IssuesEvent
2019-06-25 06:13:23
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ebay.com - site is not usable
browser-firefox-mobile engine-gecko priority-critical
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.ebay.com/i/303172994264?chn=ps&ul_ref=https%253A%252F%252Frover.ebay.com%252Frover%252F1%252F711-117182-37290-0%252F2%253Fmpre%253Dhttps%253A%252F%252Fwww.ebay.com%252Fi%252F303172994264%253Fchn%253Dps%2526var%253D602132596968%2526mkevt%253D1%2526mkrid%253D711-117182-37290-0%2526mkcid%253D2%2526itemid%253D602132596968_303172994264%2526targetid%253D503482154844%2526device%253Dm%2526adtype%253Dpla%2526googleloc%253D9029718%2526poi%253D%2526campaignid%253D1669934822%2526adgroupid%253D65058350059%2526rlsatarget%253Dpla-503482154844%2526abcId%253D1139296%2526merchantid%253D6296724%2526gclid%253DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%2526srcrot%253D711-117182-37290-0%2526rvr_id%253D2017044543838%2526rvr_ts%253D80b9d7e616b0aa48ddd4da87fff0994b **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: auto direct to ebay://link?... **Steps to Reproduce**: Even in desktop mode whole page forwarded to ebay://link?nav=item.view&id=303172994264&referrer=https%3A%2F%2Frover.ebay.com%2Frover%2F1%2F711-117182-37290-0%2F2%3Fvar%3D602132596968%26mkevt%3D1%26mkrid%3D711-117182-37290-0%26mkcid%3D2%26itemid%3D602132596968_303172994264%26targetid%3D503482154844%26device%3Dm%26adtype%3Dpla%26googleloc%3D9029718%26poi%3D%26campaignid%3D1669934822%26adgroupid%3D65058350059%26rlsatarget%3Dpla-503482154844%26abcId%3D1139296%26merchantid%3D6296724%26gclid%3DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%26srcrot%3D711-117182-37290-0%26rvr_id%3D2017044543838%26rvr_ts%3D80b9d7e616b0aa48ddd4da87fff0994b I don't have the app installed, so it doesn't load. <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190617221408</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>hasFastClick: true</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ebay.com - site is not usable - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.ebay.com/i/303172994264?chn=ps&ul_ref=https%253A%252F%252Frover.ebay.com%252Frover%252F1%252F711-117182-37290-0%252F2%253Fmpre%253Dhttps%253A%252F%252Fwww.ebay.com%252Fi%252F303172994264%253Fchn%253Dps%2526var%253D602132596968%2526mkevt%253D1%2526mkrid%253D711-117182-37290-0%2526mkcid%253D2%2526itemid%253D602132596968_303172994264%2526targetid%253D503482154844%2526device%253Dm%2526adtype%253Dpla%2526googleloc%253D9029718%2526poi%253D%2526campaignid%253D1669934822%2526adgroupid%253D65058350059%2526rlsatarget%253Dpla-503482154844%2526abcId%253D1139296%2526merchantid%253D6296724%2526gclid%253DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%2526srcrot%253D711-117182-37290-0%2526rvr_id%253D2017044543838%2526rvr_ts%253D80b9d7e616b0aa48ddd4da87fff0994b **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: auto direct to ebay://link?... **Steps to Reproduce**: Even in desktop mode whole page forwarded to ebay://link?nav=item.view&id=303172994264&referrer=https%3A%2F%2Frover.ebay.com%2Frover%2F1%2F711-117182-37290-0%2F2%3Fvar%3D602132596968%26mkevt%3D1%26mkrid%3D711-117182-37290-0%26mkcid%3D2%26itemid%3D602132596968_303172994264%26targetid%3D503482154844%26device%3Dm%26adtype%3Dpla%26googleloc%3D9029718%26poi%3D%26campaignid%3D1669934822%26adgroupid%3D65058350059%26rlsatarget%3Dpla-503482154844%26abcId%3D1139296%26merchantid%3D6296724%26gclid%3DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%26srcrot%3D711-117182-37290-0%26rvr_id%3D2017044543838%26rvr_ts%3D80b9d7e616b0aa48ddd4da87fff0994b I don't have the app installed, so it doesn't load. <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190617221408</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>hasFastClick: true</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description auto direct to ebay link steps to reproduce even in desktop mode whole page forwarded to ebay link nav item view id referrer https ebay com wcb id ts i don t have the app installed so it doesn t load browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false hasfastclick true gfx webrender enabled false gfx webrender all false channel beta from with ❤️
0
20,589
27,252,294,243
IssuesEvent
2023-02-22 09:03:27
haddocking/haddock3
https://api.github.com/repos/haddocking/haddock3
closed
add a postprocessing key to execute haddock3-analyse at the end of the workflow
analysis/postprocessing
For the workflow builder it would be nice to have `haddock3-analyse` run always at the end of the workflow. With a new option in the .cfg workflow (`postprocessing = true` ? ), haddock3 should run the postprocessing on a certain subset of folders (I would say only on the `caprieval` folders). In the workflow builder this option will be always active, while standard haddock command line users will have it set to false by default. Discussed with @sverhoeven
1.0
add a postprocessing key to execute haddock3-analyse at the end of the workflow - For the workflow builder it would be nice to have `haddock3-analyse` run always at the end of the workflow. With a new option in the .cfg workflow (`postprocessing = true` ? ), haddock3 should run the postprocessing on a certain subset of folders (I would say only on the `caprieval` folders). In the workflow builder this option will be always active, while standard haddock command line users will have it set to false by default. Discussed with @sverhoeven
process
add a postprocessing key to execute analyse at the end of the workflow for the workflow builder it would be nice to have analyse run always at the end of the workflow with a new option in the cfg workflow postprocessing true should run the postprocessing on a certain subset of folders i would say only on the caprieval folders in the workflow builder this option will be always active while standard haddock command line users will have it set to false by default discussed with sverhoeven
1
1,255
3,789,756,707
IssuesEvent
2016-03-21 19:01:25
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
opened
ProcessTests.TestExitTime failed on OSX
System.Diagnostics.Process
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/osx_debug_tst_prtest/2237/console ``` 11:35:29 System.Diagnostics.Tests.ProcessTests.TestExitTime [FAIL] 11:35:29 Assert+WrapperXunitException : File path: d:\j\workspace\osx_debug_bld---2f4069d6\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 164 11:35:29 ---- Assert.True() Failure 11:35:29 Expected: True 11:35:29 Actual: False 11:35:29 Stack Trace: 11:35:29 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber) 11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) 11:35:29 at System.Diagnostics.Tests.ProcessTests.TestExitTime() 11:35:29 ----- Inner Stack Trace ----- 11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) ```
1.0
ProcessTests.TestExitTime failed on OSX - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/osx_debug_tst_prtest/2237/console ``` 11:35:29 System.Diagnostics.Tests.ProcessTests.TestExitTime [FAIL] 11:35:29 Assert+WrapperXunitException : File path: d:\j\workspace\osx_debug_bld---2f4069d6\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 164 11:35:29 ---- Assert.True() Failure 11:35:29 Expected: True 11:35:29 Actual: False 11:35:29 Stack Trace: 11:35:29 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber) 11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) 11:35:29 at System.Diagnostics.Tests.ProcessTests.TestExitTime() 11:35:29 ----- Inner Stack Trace ----- 11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) ```
process
processtests testexittime failed on osx system diagnostics tests processtests testexittime assert wrapperxunitexception file path d j workspace osx debug bld src system diagnostics process tests processtests cs line assert true failure expected true actual false stack trace at assert wrapexception exception e string callerfilepath callerlinenumber at assert true boolean condition string usermessage string path line at system diagnostics tests processtests testexittime inner stack trace at assert true boolean condition string usermessage string path line
1
1,297
3,837,908,645
IssuesEvent
2016-04-02 02:10:17
BayoAdejare/Leverage
https://api.github.com/repos/BayoAdejare/Leverage
opened
Create additional columns to cater testing and etc
process/administration
Create additional columns to cater testing and etc
1.0
Create additional columns to cater testing and etc - Create additional columns to cater testing and etc
process
create additional columns to cater testing and etc create additional columns to cater testing and etc
1
73,648
15,278,300,224
IssuesEvent
2021-02-23 01:09:35
emilwareus/thimble.mozilla.org
https://api.github.com/repos/emilwareus/thimble.mozilla.org
opened
CVE-2020-28499 (Medium) detected in merge-1.2.1.tgz
security vulnerability
## CVE-2020-28499 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: thimble.mozilla.org/services/id.webmaker.org/package.json</p> <p>Path to vulnerable library: thimble.mozilla.org/services/id.webmaker.org/node_modules/merge/package.json</p> <p> Dependency Hierarchy: - autoless-0.1.7.tgz (Root Library) - watch-0.14.0.tgz - exec-sh-0.2.2.tgz - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/emilwareus/thimble.mozilla.org/commit/af3d91f99628f029ddcc04f2c30b6bf019be57d7">af3d91f99628f029ddcc04f2c30b6bf019be57d7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge . <p>Publish Date: 2021-02-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28499 (Medium) detected in merge-1.2.1.tgz - ## CVE-2020-28499 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: thimble.mozilla.org/services/id.webmaker.org/package.json</p> <p>Path to vulnerable library: thimble.mozilla.org/services/id.webmaker.org/node_modules/merge/package.json</p> <p> Dependency Hierarchy: - autoless-0.1.7.tgz (Root Library) - watch-0.14.0.tgz - exec-sh-0.2.2.tgz - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/emilwareus/thimble.mozilla.org/commit/af3d91f99628f029ddcc04f2c30b6bf019be57d7">af3d91f99628f029ddcc04f2c30b6bf019be57d7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge . <p>Publish Date: 2021-02-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in merge tgz cve medium severity vulnerability vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file thimble mozilla org services id webmaker org package json path to vulnerable library thimble mozilla org services id webmaker org node modules merge package json dependency hierarchy autoless tgz root library watch tgz exec sh tgz x merge tgz vulnerable library found in head commit a href vulnerability details all versions of package merge are vulnerable to prototype pollution via recursivemerge publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href step up your open source security game with whitesource
0
42,757
22,824,582,071
IssuesEvent
2022-07-12 07:24:59
Azure/azure-sdk-tools
https://api.github.com/repos/Azure/azure-sdk-tools
opened
[Perf] Move perf.yml into eng/common
pillar-performance
Most of `/eng/pipelines/templates/jobs/perf.yml` is duplicated across languages, and should be refactored into `/eng/common`.
True
[Perf] Move perf.yml into eng/common - Most of `/eng/pipelines/templates/jobs/perf.yml` is duplicated across languages, and should be refactored into `/eng/common`.
non_process
move perf yml into eng common most of eng pipelines templates jobs perf yml is duplicated across languages and should be refactored into eng common
0
18,229
24,296,427,727
IssuesEvent
2022-09-29 10:25:51
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
closed
macOS Dependencies
test-required issue-processing-state-01
Dependencies error during the installation of the quark-engine on macOS Catalina - 10.15.7. Error : - pkg_resources.DistributionNotFound: The 'androguard==3.4.0a1' distribution was not found and is required by quark-engine <img width="1016" alt="Screenshot 2020-11-02 at 6 18 56 PM" src="https://user-images.githubusercontent.com/18115054/97869748-fbcf3680-1d37-11eb-82e5-9132b3d03782.png">
1.0
macOS Dependencies - Dependencies error during the installation of the quark-engine on macOS Catalina - 10.15.7. Error : - pkg_resources.DistributionNotFound: The 'androguard==3.4.0a1' distribution was not found and is required by quark-engine <img width="1016" alt="Screenshot 2020-11-02 at 6 18 56 PM" src="https://user-images.githubusercontent.com/18115054/97869748-fbcf3680-1d37-11eb-82e5-9132b3d03782.png">
process
macos dependencies dependencies error during the installation of the quark engine on macos catalina error pkg resources distributionnotfound the androguard distribution was not found and is required by quark engine img width alt screenshot at pm src
1
7,997
11,188,128,693
IssuesEvent
2020-01-02 02:55:43
52ABP/Documents
https://api.github.com/repos/52ABP/Documents
opened
ASP.NET Core 进程外(out-of-process)托管 | 52ABP官方技术文档与博客
ASP.NET Core 进程外(out-of-process)托管 | 52ABP官方技术文档与博 Gitalk
https://docs.52abp.com/mvc/7-Out-of-ProcessHosting.html 旨在打造新手小白从入门到实战的学习网站,内容涵盖:ASP.NET Core、Angular 、.NET Core、52ABP、等企业级解决方案等
1.0
ASP.NET Core 进程外(out-of-process)托管 | 52ABP官方技术文档与博客 - https://docs.52abp.com/mvc/7-Out-of-ProcessHosting.html 旨在打造新手小白从入门到实战的学习网站,内容涵盖:ASP.NET Core、Angular 、.NET Core、52ABP、等企业级解决方案等
process
asp net core 进程外 out of process 托管 旨在打造新手小白从入门到实战的学习网站,内容涵盖:asp net core、angular 、 net core、 、等企业级解决方案等
1
66,188
8,887,044,698
IssuesEvent
2019-01-15 03:31:04
NethermindEth/nethermind
https://api.github.com/repos/NethermindEth/nethermind
closed
Ropsten sync issues / discussion
documentation
Hi Tomasz, im trying to run the Runner from within visual studio 2017. Im getting this error message: _2018-11-06 09:41:48.7958|ERROR|1|Runner.Program|Failure|Config file does not exist configs\ropsten_windows_discovery.config.json_ Are the discovery files missing from the repo? Thanks, Anders.
1.0
Ropsten sync issues / discussion - Hi Tomasz, im trying to run the Runner from within visual studio 2017. Im getting this error message: _2018-11-06 09:41:48.7958|ERROR|1|Runner.Program|Failure|Config file does not exist configs\ropsten_windows_discovery.config.json_ Are the discovery files missing from the repo? Thanks, Anders.
non_process
ropsten sync issues discussion hi tomasz im trying to run the runner from within visual studio im getting this error message error runner program failure config file does not exist configs ropsten windows discovery config json are the discovery files missing from the repo thanks anders
0
288,298
24,897,082,037
IssuesEvent
2022-10-28 16:50:07
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts - Reporting APIs Usage from new jobs posted should handle preserve_layout pdf
Feature:Reporting failed-test Team:Global Experience
A test failed on a tracked branch ``` Error: expected 500 to equal 200 at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11) at Assertion.be.Assertion.equal (node_modules/@kbn/expect/expect.js:227:8) at Assertion.be (node_modules/@kbn/expect/expect.js:69:22) at Object.waitForJobToFinish (x-pack/test/reporting_api_integration/services/usage.ts:55:31) at processTicksAndRejections (node:internal/process/task_queues:96:5) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-f408bbae8ca11f13/elastic/kibana-on-merge/kibana/x-pack/test/reporting_api_integration/services/usage.ts:98:11 at async Promise.all (index 1) at Object.expectAllJobsToFinishSuccessfully (x-pack/test/reporting_api_integration/services/usage.ts:96:7) at Context.<anonymous> (x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs.ts:50:7) at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [CI Build - 8.3](https://buildkite.com/elastic/kibana-on-merge/builds/19791#01828ea3-05b6-4d80-87c9-4b633ecd8816) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts","test.name":"Reporting APIs Usage from new jobs posted should handle preserve_layout pdf","test.failCount":1}} -->
1.0
Failing test: X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts - Reporting APIs Usage from new jobs posted should handle preserve_layout pdf - A test failed on a tracked branch ``` Error: expected 500 to equal 200 at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11) at Assertion.be.Assertion.equal (node_modules/@kbn/expect/expect.js:227:8) at Assertion.be (node_modules/@kbn/expect/expect.js:69:22) at Object.waitForJobToFinish (x-pack/test/reporting_api_integration/services/usage.ts:55:31) at processTicksAndRejections (node:internal/process/task_queues:96:5) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-f408bbae8ca11f13/elastic/kibana-on-merge/kibana/x-pack/test/reporting_api_integration/services/usage.ts:98:11 at async Promise.all (index 1) at Object.expectAllJobsToFinishSuccessfully (x-pack/test/reporting_api_integration/services/usage.ts:96:7) at Context.<anonymous> (x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs.ts:50:7) at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [CI Build - 8.3](https://buildkite.com/elastic/kibana-on-merge/builds/19791#01828ea3-05b6-4d80-87c9-4b633ecd8816) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts","test.name":"Reporting APIs Usage from new jobs posted should handle preserve_layout pdf","test.failCount":1}} -->
non_process
failing test x pack reporting api integration tests x pack test reporting api integration reporting and security usage new jobs·ts reporting apis usage from new jobs posted should handle preserve layout pdf a test failed on a tracked branch error expected to equal at assertion assert node modules kbn expect expect js at assertion be assertion equal node modules kbn expect expect js at assertion be node modules kbn expect expect js at object waitforjobtofinish x pack test reporting api integration services usage ts at processticksandrejections node internal process task queues at var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack test reporting api integration services usage ts at async promise all index at object expectalljobstofinishsuccessfully x pack test reporting api integration services usage ts at context x pack test reporting api integration reporting and security usage new jobs ts at object apply node modules kbn test target node functional test runner lib mocha wrap function js first failure
0
587,065
17,603,669,756
IssuesEvent
2021-08-17 14:37:27
kirbydesign/designsystem
https://api.github.com/repos/kirbydesign/designsystem
closed
[Enhancement] Only keep 1 destructive button (Filled)
enhancement component:Button workaround possible priority 1 stale
**Is your enhancement request related to a problem? Please describe.** In the use-cases we have right now we only have a filled destruction button (red background, black text) We might also have an outline version that has a red outline with a black text, but thats it for now. The problem with a red text is that it is not accessible. see also: https://zpl.io/VDKnZAl
1.0
[Enhancement] Only keep 1 destructive button (Filled) - **Is your enhancement request related to a problem? Please describe.** In the use-cases we have right now we only have a filled destruction button (red background, black text) We might also have an outline version that has a red outline with a black text, but thats it for now. The problem with a red text is that it is not accessible. see also: https://zpl.io/VDKnZAl
non_process
only keep destructive button filled is your enhancement request related to a problem please describe in the use cases we have right now we only have a filled destruction button red background black text we might also have an outline version that has a red outline with a black text but thats it for now the problem with a red text is that it is not accessible see also
0
560
3,021,629,256
IssuesEvent
2015-07-31 15:45:43
18F/CMS.gov-developer
https://api.github.com/repos/18F/CMS.gov-developer
opened
Plan training dates
process
* [ ] CMS to consider what kind of trainings and with you. * [ ] CMS to pick tentative dates and times.
1.0
Plan training dates - * [ ] CMS to consider what kind of trainings and with you. * [ ] CMS to pick tentative dates and times.
process
plan training dates cms to consider what kind of trainings and with you cms to pick tentative dates and times
1
12,220
14,743,131,475
IssuesEvent
2021-01-07 13:26:54
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Towne - Invoice changes?
anc-process anp-1.5 ant-support
In GitLab by @kdjstudios on Jul 16, 2019, 09:42 **Submitted by:** Deb Crown <dcrown@towneanswering.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8806788 **Server:** External **Client/Site:** Towne **Account:** NA **Issue:** I guess there was some work done behind the scenes that created a situation that – instead of printing Late Fee as a line item on the invoice – it now simply prints 1.50% Also, SA generated, printed and/or emailed invoices that have a zero balance. I don’t think it did that before, did it?
1.0
Towne - Invoice changes? - In GitLab by @kdjstudios on Jul 16, 2019, 09:42 **Submitted by:** Deb Crown <dcrown@towneanswering.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8806788 **Server:** External **Client/Site:** Towne **Account:** NA **Issue:** I guess there was some work done behind the scenes that created a situation that – instead of printing Late Fee as a line item on the invoice – it now simply prints 1.50% Also, SA generated, printed and/or emailed invoices that have a zero balance. I don’t think it did that before, did it?
process
towne invoice changes in gitlab by kdjstudios on jul submitted by deb crown helpdesk server external client site towne account na issue i guess there was some work done behind the scenes that created a situation that – instead of printing late fee as a line item on the invoice – it now simply prints also sa generated printed and or emailed invoices that have a zero balance i don’t think it did that before did it
1
5,410
8,245,815,590
IssuesEvent
2018-09-11 10:52:57
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[needs-docs][processing] return empty layer for save selected features algorithm if no features are selected
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/0cdabb5e60bd71941b559bf4f74e75e423b56536 by nirvn Unfortunately this naughty coder did not write a description... :-(
1.0
[needs-docs][processing] return empty layer for save selected features algorithm if no features are selected - Original commit: https://github.com/qgis/QGIS/commit/0cdabb5e60bd71941b559bf4f74e75e423b56536 by nirvn Unfortunately this naughty coder did not write a description... :-(
process
return empty layer for save selected features algorithm if no features are selected original commit by nirvn unfortunately this naughty coder did not write a description
1
9,569
12,519,746,233
IssuesEvent
2020-06-03 14:52:31
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
opened
[Monitoring] Skip some of the tests because they are doing cuncurrent modifications that fail.
api: monitoring priority: p1 type: process
Even thouhg in #1067 each test project was configured to not run tests in parallel, the different projects can still be being run in parallel and also we are running tests in Windows and Linux, which run roughly at the same time on presubmit. Several of the test projects are using the same fixture, and there is where we get "Too many concurrent edits to the project configuration". For fixing this we'll need to: - Merge all test projects into one and use the fixture as a collection fixture. - Add a retry on the fixture when creating the resources in case Windows and Linux are running at the same time. Tests will be disabled in #1067 until these fixes can be made.
1.0
[Monitoring] Skip some of the tests because they are doing cuncurrent modifications that fail. - Even thouhg in #1067 each test project was configured to not run tests in parallel, the different projects can still be being run in parallel and also we are running tests in Windows and Linux, which run roughly at the same time on presubmit. Several of the test projects are using the same fixture, and there is where we get "Too many concurrent edits to the project configuration". For fixing this we'll need to: - Merge all test projects into one and use the fixture as a collection fixture. - Add a retry on the fixture when creating the resources in case Windows and Linux are running at the same time. Tests will be disabled in #1067 until these fixes can be made.
process
skip some of the tests because they are doing cuncurrent modifications that fail even thouhg in each test project was configured to not run tests in parallel the different projects can still be being run in parallel and also we are running tests in windows and linux which run roughly at the same time on presubmit several of the test projects are using the same fixture and there is where we get too many concurrent edits to the project configuration for fixing this we ll need to merge all test projects into one and use the fixture as a collection fixture add a retry on the fixture when creating the resources in case windows and linux are running at the same time tests will be disabled in until these fixes can be made
1
490,066
14,115,102,726
IssuesEvent
2020-11-07 19:09:54
HHS81/c182s
https://api.github.com/repos/HHS81/c182s
closed
use `/fdm/jsbsim/propulsion/engine/AFR` for exhaust smoke in rich operation
effects low priority refactoring wontfix
`/fdm/jsbsim/propulsion/engine/AFR` offers the Air-to-Fuel ratio in the current FGFS GIT version. This could be used to generate smoke from the exhaust when mixture is very rich and may also replace the naive start-smoke implementation from #297 (which occurs also because of very rich starter mixture).
1.0
use `/fdm/jsbsim/propulsion/engine/AFR` for exhaust smoke in rich operation - `/fdm/jsbsim/propulsion/engine/AFR` offers the Air-to-Fuel ratio in the current FGFS GIT version. This could be used to generate smoke from the exhaust when mixture is very rich and may also replace the naive start-smoke implementation from #297 (which occurs also because of very rich starter mixture).
non_process
use fdm jsbsim propulsion engine afr for exhaust smoke in rich operation fdm jsbsim propulsion engine afr offers the air to fuel ratio in the current fgfs git version this could be used to generate smoke from the exhaust when mixture is very rich and may also replace the naive start smoke implementation from which occurs also because of very rich starter mixture
0
284,091
8,735,673,193
IssuesEvent
2018-12-11 17:22:02
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
Gui crashes attempting to open on remote after using 'save session' dialog
bug likelihood medium priority reviewed severity high
Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2385 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: Gui crashes attempting to open on remote after using 'save session' dialog Assigned to: Category: Target version: 2.10 Author: Kathleen Biagas Start: 09/22/2015 Due date: % Done: 0 Estimated time: Created: 09/22/2015 12:50 pm Updated: 10/06/2015 08:40 pm Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: trunk Impact: Expected Use: OS: All Support Group: Any Description: Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash. Comments: I tried to duplicate this problem on a Linux system, but could not duplicate the crash.But I am not sure if I really duplicated the problem because I have questions about the steps.Also Brad Whitlock made some changes to the session save code and this could have fixed the issue, but here are the steps I tried to duplicate the problem.1) Open a data file on the local machine. I opened the noise data set.2) Created a Mesh plot on the mesh variable.3) Open the "Save session" dialog and then clicked cancel.4) Here is were I am not sure what "Open" means, which dialog "Save session", "Restore session", or "Open file". I am also not sure what "select a remote host" means, does that mean clicked on the host drop down or field itself? So I tried the following: 1) Open the "Save session" dialog. Then click on the host field text box and selected the drop down. 2) Open the "Restore session" dialog. Then click on the host field text box and selected the drop down. 3) Open the "Open file" dialog. Then click on the host field text box and selected the drop down.None of these steps caused a crash. So did I miss something in duplicating the problem?Thanks,David 1) Open a file and draw a plot2) Open 'Save session' dialog, cancel it3) Click 'Open' to get the file open dialog, select the host dropdown and choose a remote hostGui will crash.gdb shows a segv from QvisFileWindowBase::ProgressCallback in the call toThis->setEnabled(false) Short way to duplicate the problem:1) Click the "Open" data button on GUI and click "Cancel" button.2) Click File->"Save session" to open dialog and click "Cancel" button.3) Click the "Open" data button on GUI and select different host in the host drop down.Crash inside Qt setEnabled_helper called by QvisFileWindowBase::ProgressCallback.So I think there must be something shared between the Open Data Window and the Session save dialog and when the Session Save dialog is deleted it must close or delete something they both share. Resolved by David Camp with SVN #27306, which was ported to 2.10RC with SVN #27402.
1.0
Gui crashes attempting to open on remote after using 'save session' dialog - Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2385 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: Gui crashes attempting to open on remote after using 'save session' dialog Assigned to: Category: Target version: 2.10 Author: Kathleen Biagas Start: 09/22/2015 Due date: % Done: 0 Estimated time: Created: 09/22/2015 12:50 pm Updated: 10/06/2015 08:40 pm Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: trunk Impact: Expected Use: OS: All Support Group: Any Description: Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash. Comments: I tried to duplicate this problem on a Linux system, but could not duplicate the crash.But I am not sure if I really duplicated the problem because I have questions about the steps.Also Brad Whitlock made some changes to the session save code and this could have fixed the issue, but here are the steps I tried to duplicate the problem.1) Open a data file on the local machine. I opened the noise data set.2) Created a Mesh plot on the mesh variable.3) Open the "Save session" dialog and then clicked cancel.4) Here is were I am not sure what "Open" means, which dialog "Save session", "Restore session", or "Open file". I am also not sure what "select a remote host" means, does that mean clicked on the host drop down or field itself? So I tried the following: 1) Open the "Save session" dialog. Then click on the host field text box and selected the drop down. 2) Open the "Restore session" dialog. Then click on the host field text box and selected the drop down. 3) Open the "Open file" dialog. Then click on the host field text box and selected the drop down.None of these steps caused a crash. So did I miss something in duplicating the problem?Thanks,David 1) Open a file and draw a plot2) Open 'Save session' dialog, cancel it3) Click 'Open' to get the file open dialog, select the host dropdown and choose a remote hostGui will crash.gdb shows a segv from QvisFileWindowBase::ProgressCallback in the call toThis->setEnabled(false) Short way to duplicate the problem:1) Click the "Open" data button on GUI and click "Cancel" button.2) Click File->"Save session" to open dialog and click "Cancel" button.3) Click the "Open" data button on GUI and select different host in the host drop down.Crash inside Qt setEnabled_helper called by QvisFileWindowBase::ProgressCallback.So I think there must be something shared between the Open Data Window and the Session save dialog and when the Session Save dialog is deleted it must close or delete something they both share. Resolved by David Camp with SVN #27306, which was ported to 2.10RC with SVN #27402.
non_process
gui crashes attempting to open on remote after using save session dialog open a file and draw a plot on the local machine open save session dialog but don t do anything just cancel open select a remote host the gui will crash redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject gui crashes attempting to open on remote after using save session dialog assigned to category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version trunk impact expected use os all support group any description open a file and draw a plot on the local machine open save session dialog but don t do anything just cancel open select a remote host the gui will crash comments i tried to duplicate this problem on a linux system but could not duplicate the crash but i am not sure if i really duplicated the problem because i have questions about the steps also brad whitlock made some changes to the session save code and this could have fixed the issue but here are the steps i tried to duplicate the problem open a data file on the local machine i opened the noise data set created a mesh plot on the mesh variable open the save session dialog and then clicked cancel here is were i am not sure what open means which dialog save session restore session or open file i am also not sure what select a remote host means does that mean clicked on the host drop down or field itself so i tried the following open the save session dialog then click on the host field text box and selected the drop down open the restore session dialog then click on the host field text box and selected the drop down open the open file dialog then click on the host field text box and selected the drop down none of these steps caused a crash so did i miss something in duplicating the problem thanks david open a file and draw a open save session dialog cancel click open to get the file open dialog select the host dropdown and choose a remote hostgui will crash gdb shows a segv from qvisfilewindowbase progresscallback in the call tothis setenabled false short way to duplicate the problem click the open data button on gui and click cancel button click file save session to open dialog and click cancel button click the open data button on gui and select different host in the host drop down crash inside qt setenabled helper called by qvisfilewindowbase progresscallback so i think there must be something shared between the open data window and the session save dialog and when the session save dialog is deleted it must close or delete something they both share resolved by david camp with svn which was ported to with svn
0
224,674
24,783,421,931
IssuesEvent
2022-10-24 07:50:21
sast-automation-dev/openidm-community-edition-43
https://api.github.com/repos/sast-automation-dev/openidm-community-edition-43
opened
openidm-config-2.1.2.jar: 1 vulnerabilities (highest severity is: 6.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openidm-config-2.1.2.jar</b></p></summary> <p></p> <p>Path to dependency file: /openidm-shell/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (openidm-config version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2019-0379](https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | commons-codec-1.7.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0379</summary> ### Vulnerable Library - <b>commons-codec-1.7.jar</b></p> <p>The codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Library home page: <a href="http://commons.apache.org/codec/">http://commons.apache.org/codec/</a></p> <p>Path to dependency file: /openidm-repo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p> <p> Dependency Hierarchy: - openidm-config-2.1.2.jar (Root Library) - openidm-repo-2.1.2.jar - openidm-util-2.1.2.jar - json-crypto-core-1.1.0.jar - :x: **commons-codec-1.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: May 20, 2019 3:39:18 PM <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: May 20, 2019 3:39:18 PM</p> <p>Fix Resolution: commons-codec:commons-codec:1.13</p> </p> <p></p> </details>
True
openidm-config-2.1.2.jar: 1 vulnerabilities (highest severity is: 6.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openidm-config-2.1.2.jar</b></p></summary> <p></p> <p>Path to dependency file: /openidm-shell/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (openidm-config version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2019-0379](https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | commons-codec-1.7.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0379</summary> ### Vulnerable Library - <b>commons-codec-1.7.jar</b></p> <p>The codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Library home page: <a href="http://commons.apache.org/codec/">http://commons.apache.org/codec/</a></p> <p>Path to dependency file: /openidm-repo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p> <p> Dependency Hierarchy: - openidm-config-2.1.2.jar (Root Library) - openidm-repo-2.1.2.jar - openidm-util-2.1.2.jar - json-crypto-core-1.1.0.jar - :x: **commons-codec-1.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: May 20, 2019 3:39:18 PM <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: May 20, 2019 3:39:18 PM</p> <p>Fix Resolution: commons-codec:commons-codec:1.13</p> </p> <p></p> </details>
non_process
openidm config jar vulnerabilities highest severity is vulnerable library openidm config jar path to dependency file openidm shell pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in openidm config version remediation available medium commons codec jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details ws vulnerable library commons codec jar the codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities library home page a href path to dependency file openidm repo pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy openidm config jar root library openidm repo jar openidm util jar json crypto core jar x commons codec jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date may pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date may pm fix resolution commons codec commons codec
0
22,494
31,468,484,843
IssuesEvent
2023-08-30 05:19:51
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
When save questions, the placeholder name is not consistent
Type:Bug Priority:P3 .Backend Querying/GUI .metabase-lib .Team/QueryProcessor :hammer_and_wrench: .Wanted: MLv2
**Describe the bug** When saving a question, we automatically fill out a name for it based on the source table name. But when a question has a join, the name does not contain the joined table. **To Reproduce** 1. New Question 2. Sample Dataset -> Orders 3. Join with the Products table 4. Click save ![Screen Shot 2022-08-10 at 20 17 16](https://user-images.githubusercontent.com/25661381/183911446-d3857801-3e98-4876-a3b2-e5fd08da6e2f.png) 5. Notice the name in the form (1st arrow) is "Orders", But the name of the question (2nd arrow) is "Orders + Products" which I think is a neat auto format. **Expected behavior** The name should be "Orders + Products" **Severity** Minor **Additional context** Tested with current master(1ca62d1b43cc97dc9af4886fa0abee8aa025a47c)
1.0
When save questions, the placeholder name is not consistent - **Describe the bug** When saving a question, we automatically fill out a name for it based on the source table name. But when a question has a join, the name does not contain the joined table. **To Reproduce** 1. New Question 2. Sample Dataset -> Orders 3. Join with the Products table 4. Click save ![Screen Shot 2022-08-10 at 20 17 16](https://user-images.githubusercontent.com/25661381/183911446-d3857801-3e98-4876-a3b2-e5fd08da6e2f.png) 5. Notice the name in the form (1st arrow) is "Orders", But the name of the question (2nd arrow) is "Orders + Products" which I think is a neat auto format. **Expected behavior** The name should be "Orders + Products" **Severity** Minor **Additional context** Tested with current master(1ca62d1b43cc97dc9af4886fa0abee8aa025a47c)
process
when save questions the placeholder name is not consistent describe the bug when saving a question we automatically fill out a name for it based on the source table name but when a question has a join the name does not contain the joined table to reproduce new question sample dataset orders join with the products table click save notice the name in the form arrow is orders but the name of the question arrow is orders products which i think is a neat auto format expected behavior the name should be orders products severity minor additional context tested with current master
1
83,381
3,634,281,793
IssuesEvent
2016-02-11 17:24:59
PCMDI/pcmdi_metrics
https://api.github.com/repos/PCMDI/pcmdi_metrics
reopened
output clim not masked for ocean case...
bug TOPPRIORITY
In the parameter file, we can select variables for which we want metrics on 'global', 'ocean' and 'land' case (e.g., below). Global and ocean cases are ok, but the output.nc clims for the 'ocean' case are not masked over land. ## REGIONS ON WHICH WE WANT TO RUN METRICS (var specific) regions = {"tas" : [None,"land","ocean"], "uas" : [None,"land","ocean"], "vas" : [None,"land","ocean"], "pr" : [None,"land","ocean"], "psl": [None,"land","ocean",], "huss": [None,"land","ocean"], "prw": [None,"land","ocean"],
1.0
output clim not masked for ocean case... - In the parameter file, we can select variables for which we want metrics on 'global', 'ocean' and 'land' case (e.g., below). Global and ocean cases are ok, but the output.nc clims for the 'ocean' case are not masked over land. ## REGIONS ON WHICH WE WANT TO RUN METRICS (var specific) regions = {"tas" : [None,"land","ocean"], "uas" : [None,"land","ocean"], "vas" : [None,"land","ocean"], "pr" : [None,"land","ocean"], "psl": [None,"land","ocean",], "huss": [None,"land","ocean"], "prw": [None,"land","ocean"],
non_process
output clim not masked for ocean case in the parameter file we can select variables for which we want metrics on global ocean and land case e g below global and ocean cases are ok but the output nc clims for the ocean case are not masked over land regions on which we want to run metrics var specific regions tas uas vas pr psl huss prw
0
175,817
27,979,244,304
IssuesEvent
2023-03-26 00:16:28
kocheck/Lazy-Export
https://api.github.com/repos/kocheck/Lazy-Export
closed
New Branding
Needs Design no-issue-activity
## Overview Let's be honest, we need a logo and Artwork created. I have been putting this off and I really should dig deep here for something fun.
1.0
New Branding - ## Overview Let's be honest, we need a logo and Artwork created. I have been putting this off and I really should dig deep here for something fun.
non_process
new branding overview let s be honest we need a logo and artwork created i have been putting this off and i really should dig deep here for something fun
0
7,405
10,523,941,063
IssuesEvent
2019-09-30 12:15:14
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
Lift broken without init flow
process/candidate
```sh $ ./node_modules/.bin/prisma2 introspect --pg-host localhost --pg-password "" --pg-user m --pg-schema public --pg-db prisma-blog $ mv datamodel* schema.prisma $ ./node_modules/.bin/prisma2 lift up ERROR Oops, an unexpected error occured! Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.: Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5 stack backtrace: 0: std::panicking::default_hook::{{closure}} ERROR Oops, an unexpected error occured! Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.: Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5 stack backtrace: 0: std::panicking::default_hook::{{closure}} 1: std::panicking::default_hook 2: migration_engine::main::{{closure}} 3: std::panicking::rust_panic_with_hook 4: std::panicking::continue_panic_fmt 5: rust_begin_unwind 6: core::panicking::panic_fmt 7: core::result::unwrap_failed 8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::deserialize_databas e_migration 9: <migration_engine::commands::list_migrations::ListMigrationStepsCommand as migration_engine::commands::command:: MigrationCommand>::execute 10: <migration_engine::api::MigrationApi<C,D> as migration_engine::api::GenericApi>::list_migrations 11: migration_engine::api::rpc::RpcApi::create_sync_handler 12: tokio_executor::enter::exit 13: tokio_threadpool::blocking::blocking 14: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll 15: futures::future::chain::Chain<A,B,C>::poll 16: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll 17: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll 18: futures::future::chain::Chain<A,B,C>::poll 19: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll 20: <futures::future::map::Map<A,F> as futures::future::Future>::poll 21: <futures::future::either::Either<A,B> as futures::future::Future>::poll 22: <futures::future::map::Map<A,F> as futures::future::Future>::poll 23: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll 24: <futures::stream::and_then::AndThen<S,F,U> as futures::stream::Stream>::poll 25: <futures::stream::forward::Forward<T,U> as futures::future::Future>::poll 26: <futures::future::map::Map<A,F> as futures::future::Future>::poll 27: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll 28: futures::task_impl::std::set 29: std::panicking::try::do_call 30: __rust_maybe_catch_panic 31: tokio_threadpool::task::Task::run 32: tokio_threadpool::worker::Worker::run_task 33: tokio_threadpool::worker::Worker::run 34: std::thread::local::LocalKey<T>::with 35: std::thread::local::LocalKey<T>::with 36: tokio_reactor::with_default 37: tokio::runtime::threadpool::builder::Builder::build::{{closure}} 38: std::thread::local::LocalKey<T>::with 39: std::thread::local::LocalKey<T>::with note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace. Please help us improve Prisma 2 by submitting an error report. Error reports never contain personal or other sensitive information. Learn more: https://pris.ly/d/telemetry ```
1.0
Lift broken without init flow - ```sh $ ./node_modules/.bin/prisma2 introspect --pg-host localhost --pg-password "" --pg-user m --pg-schema public --pg-db prisma-blog $ mv datamodel* schema.prisma $ ./node_modules/.bin/prisma2 lift up ERROR Oops, an unexpected error occured! Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.: Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5 stack backtrace: 0: std::panicking::default_hook::{{closure}} ERROR Oops, an unexpected error occured! Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.: Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5 stack backtrace: 0: std::panicking::default_hook::{{closure}} 1: std::panicking::default_hook 2: migration_engine::main::{{closure}} 3: std::panicking::rust_panic_with_hook 4: std::panicking::continue_panic_fmt 5: rust_begin_unwind 6: core::panicking::panic_fmt 7: core::result::unwrap_failed 8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::deserialize_databas e_migration 9: <migration_engine::commands::list_migrations::ListMigrationStepsCommand as migration_engine::commands::command:: MigrationCommand>::execute 10: <migration_engine::api::MigrationApi<C,D> as migration_engine::api::GenericApi>::list_migrations 11: migration_engine::api::rpc::RpcApi::create_sync_handler 12: tokio_executor::enter::exit 13: tokio_threadpool::blocking::blocking 14: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll 15: futures::future::chain::Chain<A,B,C>::poll 16: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll 17: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll 18: futures::future::chain::Chain<A,B,C>::poll 19: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll 20: <futures::future::map::Map<A,F> as futures::future::Future>::poll 21: <futures::future::either::Either<A,B> as futures::future::Future>::poll 22: <futures::future::map::Map<A,F> as futures::future::Future>::poll 23: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll 24: <futures::stream::and_then::AndThen<S,F,U> as futures::stream::Stream>::poll 25: <futures::stream::forward::Forward<T,U> as futures::future::Future>::poll 26: <futures::future::map::Map<A,F> as futures::future::Future>::poll 27: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll 28: futures::task_impl::std::set 29: std::panicking::try::do_call 30: __rust_maybe_catch_panic 31: tokio_threadpool::task::Task::run 32: tokio_threadpool::worker::Worker::run_task 33: tokio_threadpool::worker::Worker::run 34: std::thread::local::LocalKey<T>::with 35: std::thread::local::LocalKey<T>::with 36: tokio_reactor::with_default 37: tokio::runtime::threadpool::builder::Builder::build::{{closure}} 38: std::thread::local::LocalKey<T>::with 39: std::thread::local::LocalKey<T>::with note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace. Please help us improve Prisma 2 by submitting an error report. Error reports never contain personal or other sensitive information. Learn more: https://pris.ly/d/telemetry ```
process
lift broken without init flow sh node modules bin introspect pg host localhost pg password pg user m pg schema public pg db prisma blog mv datamodel schema prisma node modules bin lift up error oops an unexpected error occured error in migration engine thread tokio runtime worker panicked at deserializing the database migration failed error missing field original steps line column src libcore result rs stack backtrace std panicking default hook closure error oops an unexpected error occured error in migration engine thread tokio runtime worker panicked at deserializing the database migration failed error missing field original steps line column src libcore result rs stack backtrace std panicking default hook closure std panicking default hook migration engine main closure std panicking rust panic with hook std panicking continue panic fmt rust begin unwind core panicking panic fmt core result unwrap failed deserialize databas e migration migration engine commands list migrations listmigrationstepscommand as migration engine commands command migrationcommand execute as migration engine api genericapi list migrations migration engine api rpc rpcapi create sync handler tokio executor enter exit tokio threadpool blocking blocking as futures future future poll futures future chain chain poll as futures future future poll as futures future future poll futures future chain chain poll as futures future future poll as futures future future poll as futures future future poll as futures future future poll as futures future future poll as futures stream stream poll as futures future future poll as futures future future poll as futures future future poll futures task impl std set std panicking try do call rust maybe catch panic tokio threadpool task task run tokio threadpool worker worker run task tokio threadpool worker worker run std thread local localkey with std thread local localkey with tokio reactor with default tokio runtime threadpool builder builder build closure std thread local localkey with std thread local localkey with note some details are omitted run with rust backtrace full for a verbose backtrace please help us improve prisma by submitting an error report error reports never contain personal or other sensitive information learn more
1
361,035
10,702,851,520
IssuesEvent
2019-10-24 08:23:38
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
opened
Analytics should be reported from the iOS app
Priority: P0 applications ios wallet
### Expected Behavior Analytics are reported from the iOS app. ### Current Behavior Analytics are not reported from the iOS app.
1.0
Analytics should be reported from the iOS app - ### Expected Behavior Analytics are reported from the iOS app. ### Current Behavior Analytics are not reported from the iOS app.
non_process
analytics should be reported from the ios app expected behavior analytics are reported from the ios app current behavior analytics are not reported from the ios app
0
10,223
13,093,700,514
IssuesEvent
2020-08-03 10:54:51
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
improvement of spatial selection
Feature Request Feedback Processing
Author Name: **jannekevandijk -** (jannekevandijk -) Original Redmine Issue: [3703](https://issues.qgis.org/issues/3703) Redmine category:c++_plugins/spatial_query --- basically something similar to [[ArcGis]] 'select by location', where you are able to: 1. select features in multiple layers 2. select using a buffer without having to create buffers first (eg select features within a distance of selected features of another layer) (or should this be two seperate tickets?)
1.0
improvement of spatial selection - Author Name: **jannekevandijk -** (jannekevandijk -) Original Redmine Issue: [3703](https://issues.qgis.org/issues/3703) Redmine category:c++_plugins/spatial_query --- basically something similar to [[ArcGis]] 'select by location', where you are able to: 1. select features in multiple layers 2. select using a buffer without having to create buffers first (eg select features within a distance of selected features of another layer) (or should this be two seperate tickets?)
process
improvement of spatial selection author name jannekevandijk jannekevandijk original redmine issue redmine category c plugins spatial query basically something similar to select by location where you are able to select features in multiple layers select using a buffer without having to create buffers first eg select features within a distance of selected features of another layer or should this be two seperate tickets
1
424,692
12,322,166,329
IssuesEvent
2020-05-13 09:53:27
ooni/probe-engine
https://api.github.com/repos/ooni/probe-engine
opened
cleanup: stop downloading CA as a resource
effort/XS priority/low
We cannot do that now, because MK code still needs it. We will be able to do so when we have removed all MK code. At that point, also the CABundlePath of the session can be removed. This follows-up from https://github.com/ooni/probe-engine/issues/335
1.0
cleanup: stop downloading CA as a resource - We cannot do that now, because MK code still needs it. We will be able to do so when we have removed all MK code. At that point, also the CABundlePath of the session can be removed. This follows-up from https://github.com/ooni/probe-engine/issues/335
non_process
cleanup stop downloading ca as a resource we cannot do that now because mk code still needs it we will be able to do so when we have removed all mk code at that point also the cabundlepath of the session can be removed this follows up from
0
15,562
19,703,503,948
IssuesEvent
2022-01-12 19:08:02
googleapis/java-security-private-ca
https://api.github.com/repos/googleapis/java-security-private-ca
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'security-private-ca' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'security-private-ca' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname security private ca invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
243,667
26,287,388,377
IssuesEvent
2023-01-08 01:04:09
Yoavmartin/spring-petclinic
https://api.github.com/repos/Yoavmartin/spring-petclinic
opened
CVE-2019-8331 (Medium) detected in bootstrap-3.3.6.jar
security vulnerability
## CVE-2019-8331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.jar</b></p></summary> <p>WebJar for Bootstrap</p> <p>Library home page: <a href="http://webjars.org">http://webjars.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /canner/.m2/repository/org/webjars/bootstrap/3.3.6/bootstrap-3.3.6.jar</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.6.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: 3.4.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2019-8331 (Medium) detected in bootstrap-3.3.6.jar - ## CVE-2019-8331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.jar</b></p></summary> <p>WebJar for Bootstrap</p> <p>Library home page: <a href="http://webjars.org">http://webjars.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /canner/.m2/repository/org/webjars/bootstrap/3.3.6/bootstrap-3.3.6.jar</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.6.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: 3.4.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_process
cve medium detected in bootstrap jar cve medium severity vulnerability vulnerable library bootstrap jar webjar for bootstrap library home page a href path to dependency file pom xml path to vulnerable library canner repository org webjars bootstrap bootstrap jar dependency hierarchy x bootstrap jar vulnerable library found in base branch main vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue
0
646,401
21,046,878,003
IssuesEvent
2022-03-31 16:50:14
Zydrax/Baldazzar-BUG-Traker
https://api.github.com/repos/Zydrax/Baldazzar-BUG-Traker
closed
Quest - Preemptive Methods (26873)
Medium Priority
Current behaviour: Serpentbloom has no loot Expected behaviour: Serpentbloom should drop 1 Serpentbloom (26873) Steps to reproduce the problem: Loot Include proofs for this behaviour https://www.wowhead.com/quest=26873/preemptive-methods
1.0
Quest - Preemptive Methods (26873) - Current behaviour: Serpentbloom has no loot Expected behaviour: Serpentbloom should drop 1 Serpentbloom (26873) Steps to reproduce the problem: Loot Include proofs for this behaviour https://www.wowhead.com/quest=26873/preemptive-methods
non_process
quest preemptive methods current behaviour serpentbloom has no loot expected behaviour serpentbloom should drop serpentbloom steps to reproduce the problem loot include proofs for this behaviour
0
4,658
7,496,686,023
IssuesEvent
2018-04-08 12:11:35
ODiogoSilva/assemblerflow
https://api.github.com/repos/ODiogoSilva/assemblerflow
closed
Add assembly process template for skesa
enhancement process
This template is akin to spade's `process_spades` template that is meant to parse and assembly by contig size and average kmer coverage.
1.0
Add assembly process template for skesa - This template is akin to spade's `process_spades` template that is meant to parse and assembly by contig size and average kmer coverage.
process
add assembly process template for skesa this template is akin to spade s process spades template that is meant to parse and assembly by contig size and average kmer coverage
1
64,044
12,413,786,736
IssuesEvent
2020-05-22 13:24:09
eclipse/che
https://api.github.com/repos/eclipse/che
closed
Git clone using private key with password is not working
area/che-theia kind/bug severity/P1 status/code-review
### Describe the bug When trying to clone git repository using `git@gitlab.com/something/something` using ssh key, which is protected by password, Che never asks me for a password and the clone fails with this error: ``` git clone git@gitlab.something/something/web-app.git /projects/web-app Cloning into '/projects/web-app'... git@gitlab.com: Permission denied (publickey). fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ``` ### Che version che.openshift.io - [x] latest - [ ] nightly - [ ] other: please specify ### Steps to reproduce * Generate ssh keypair *with a password* (`ssh-keygen`) * Upload public part of that key to gitlab/github * Start any workspace * upload private part of the key to the workspace * try to clone some repo using this format -> `git@gitlab.something/something/web-app.git` ### Expected behavior Repo is cloned
1.0
Git clone using private key with password is not working - ### Describe the bug When trying to clone git repository using `git@gitlab.com/something/something` using ssh key, which is protected by password, Che never asks me for a password and the clone fails with this error: ``` git clone git@gitlab.something/something/web-app.git /projects/web-app Cloning into '/projects/web-app'... git@gitlab.com: Permission denied (publickey). fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ``` ### Che version che.openshift.io - [x] latest - [ ] nightly - [ ] other: please specify ### Steps to reproduce * Generate ssh keypair *with a password* (`ssh-keygen`) * Upload public part of that key to gitlab/github * Start any workspace * upload private part of the key to the workspace * try to clone some repo using this format -> `git@gitlab.something/something/web-app.git` ### Expected behavior Repo is cloned
non_process
git clone using private key with password is not working describe the bug when trying to clone git repository using git gitlab com something something using ssh key which is protected by password che never asks me for a password and the clone fails with this error git clone git gitlab something something web app git projects web app cloning into projects web app git gitlab com permission denied publickey fatal could not read from remote repository please make sure you have the correct access rights and the repository exists che version che openshift io latest nightly other please specify steps to reproduce generate ssh keypair with a password ssh keygen upload public part of that key to gitlab github start any workspace upload private part of the key to the workspace try to clone some repo using this format git gitlab something something web app git expected behavior repo is cloned
0
3,925
6,845,689,064
IssuesEvent
2017-11-13 09:17:46
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
closed
Redirected volume removal times out and leaves stale metadata
priority_normal process_wontfix type_bug
Scenario: - config: - redirect timeout 5 secs - backend timeout 15 secs - volume V is running on node N - SCO write W to the backend is happening in the background - volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish - W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires => unlink call times out on N
1.0
Redirected volume removal times out and leaves stale metadata - Scenario: - config: - redirect timeout 5 secs - backend timeout 15 secs - volume V is running on node N - SCO write W to the backend is happening in the background - volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish - W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires => unlink call times out on N
process
redirected volume removal times out and leaves stale metadata scenario config redirect timeout secs backend timeout secs volume v is running on node n sco write w to the backend is happening in the background volume removal is started on node m redirect to node n removal needs to wait for w to finish w encounters backend issues finishes unsuccessfully when the backend timeout fires unlink call times out on n
1
21,434
29,477,581,720
IssuesEvent
2023-06-02 00:34:05
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] [Bug] Incorrect display name for cumulative count operator
Type:Bug Querying/GUI Querying/Notebook .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
Once applied, the "Cumulative count" operator is displayed like "Cumulative count of nil" in the notebook editor This name is coming from `ML.displayInfo(query, index, aggregationClause).displayName` Can be reproduced on aggregations integration branch (#31209) ### To Reproduce 1. New > Question > Any data source 2. Click "Pick the metric you want to see" 3. Pick the "Cumulative count of rows" option in "Basic operators" **Expected:** the clause shows "Cumulative count" **Actual:** the clause shows "Cumulative count of nil" ### Demo
1.0
[MLv2] [Bug] Incorrect display name for cumulative count operator - Once applied, the "Cumulative count" operator is displayed like "Cumulative count of nil" in the notebook editor This name is coming from `ML.displayInfo(query, index, aggregationClause).displayName` Can be reproduced on aggregations integration branch (#31209) ### To Reproduce 1. New > Question > Any data source 2. Click "Pick the metric you want to see" 3. Pick the "Cumulative count of rows" option in "Basic operators" **Expected:** the clause shows "Cumulative count" **Actual:** the clause shows "Cumulative count of nil" ### Demo
process
incorrect display name for cumulative count operator once applied the cumulative count operator is displayed like cumulative count of nil in the notebook editor this name is coming from ml displayinfo query index aggregationclause displayname can be reproduced on aggregations integration branch to reproduce new question any data source click pick the metric you want to see pick the cumulative count of rows option in basic operators expected the clause shows cumulative count actual the clause shows cumulative count of nil demo
1
16,986
22,351,057,948
IssuesEvent
2022-06-15 12:06:57
python/cpython
https://api.github.com/repos/python/cpython
closed
multiprocessing with maxtasksperchild can hang if unpickling causes import
type-bug expert-multiprocessing
**Bug report** This seems like another specific instance of the general issue identified in #50970. If `multiprocessing.Pool.map_async` is used with `maxtasksperchild` and a value returned by a task is of a class not currently imported by the calling process, it can lead to a hang. Here is an example that reliably hangs for me, but which exits cleanly if ElementTree is imported at the top level. ```python #!/usr/bin/env python import os import multiprocessing def worker(num: int): from xml.etree.ElementTree import ElementTree print(f"Worker {num} with pid {os.getpid()}") return ElementTree() def main(cores: int = 4, num: int = 6): pool = multiprocessing.Pool(processes=cores, maxtasksperchild=1) barList = list(pool.map_async(worker, list(range(num))).get()) print(barList) if __name__ == "__main__": main() ``` Running `py-spy dump` on one of the workers shows this backtrace: ``` Process 47102: python ./demo_core.py Python v3.10.4 (/usr/bin/python3.10) Thread 47102 (idle): "Thread-1 (_handle_workers)" acquire (<frozen importlib._bootstrap>:120) __enter__ (<frozen importlib._bootstrap>:171) _find_and_load (<frozen importlib._bootstrap>:1024) worker (demo_core.py:8) mapstar (multiprocessing/pool.py:48) worker (multiprocessing/pool.py:125) run (multiprocessing/process.py:108) _bootstrap (multiprocessing/process.py:315) _launch (multiprocessing/popen_fork.py:71) __init__ (multiprocessing/popen_fork.py:19) _Popen (multiprocessing/context.py:277) start (multiprocessing/process.py:121) _repopulate_pool_static (multiprocessing/pool.py:326) _maintain_pool (multiprocessing/pool.py:337) _handle_workers (multiprocessing/pool.py:513) run (threading.py:946) _bootstrap_inner (threading.py:1009) _bootstrap (threading.py:966) ``` My guess (without any further proof) is that the main process receives a pickled ElementTree and starts importing the module. Concurrently, another thread realises it needs to start a new worker, so does a `fork()`. The child process has a half-imported, locked ElementTree module, and tries to import it again, leading to a deadlock. Note that this is nothing to do with ElementTree - I get the same behaviour with numpy. I chose ElementTree as a reasonably complex module (to maximise the window for the race condition) with a picklable class. Personally I consider the `fork` model of `multiprocessing` to be dangerous and requiring of care to ensure all worker tasks are created before doing anything that can conceivably create threads, and definitely a bad combination with `maxtasksperchild`. So I won't shed any tears if the resolution is "won't fix, don't do that". But https://github.com/python/cpython/issues/50970#issuecomment-1147328694 seems to suggest that @vstinner has some appetite for addressing such issues and hence I'm filing this. **Your environment** - CPython versions tested on: 3.8.10, 3.10.4 - Operating system and architecture: Ubuntu 20.04, x86_64
1.0
multiprocessing with maxtasksperchild can hang if unpickling causes import - **Bug report** This seems like another specific instance of the general issue identified in #50970. If `multiprocessing.Pool.map_async` is used with `maxtasksperchild` and a value returned by a task is of a class not currently imported by the calling process, it can lead to a hang. Here is an example that reliably hangs for me, but which exits cleanly if ElementTree is imported at the top level. ```python #!/usr/bin/env python import os import multiprocessing def worker(num: int): from xml.etree.ElementTree import ElementTree print(f"Worker {num} with pid {os.getpid()}") return ElementTree() def main(cores: int = 4, num: int = 6): pool = multiprocessing.Pool(processes=cores, maxtasksperchild=1) barList = list(pool.map_async(worker, list(range(num))).get()) print(barList) if __name__ == "__main__": main() ``` Running `py-spy dump` on one of the workers shows this backtrace: ``` Process 47102: python ./demo_core.py Python v3.10.4 (/usr/bin/python3.10) Thread 47102 (idle): "Thread-1 (_handle_workers)" acquire (<frozen importlib._bootstrap>:120) __enter__ (<frozen importlib._bootstrap>:171) _find_and_load (<frozen importlib._bootstrap>:1024) worker (demo_core.py:8) mapstar (multiprocessing/pool.py:48) worker (multiprocessing/pool.py:125) run (multiprocessing/process.py:108) _bootstrap (multiprocessing/process.py:315) _launch (multiprocessing/popen_fork.py:71) __init__ (multiprocessing/popen_fork.py:19) _Popen (multiprocessing/context.py:277) start (multiprocessing/process.py:121) _repopulate_pool_static (multiprocessing/pool.py:326) _maintain_pool (multiprocessing/pool.py:337) _handle_workers (multiprocessing/pool.py:513) run (threading.py:946) _bootstrap_inner (threading.py:1009) _bootstrap (threading.py:966) ``` My guess (without any further proof) is that the main process receives a pickled ElementTree and starts importing the module. Concurrently, another thread realises it needs to start a new worker, so does a `fork()`. The child process has a half-imported, locked ElementTree module, and tries to import it again, leading to a deadlock. Note that this is nothing to do with ElementTree - I get the same behaviour with numpy. I chose ElementTree as a reasonably complex module (to maximise the window for the race condition) with a picklable class. Personally I consider the `fork` model of `multiprocessing` to be dangerous and requiring of care to ensure all worker tasks are created before doing anything that can conceivably create threads, and definitely a bad combination with `maxtasksperchild`. So I won't shed any tears if the resolution is "won't fix, don't do that". But https://github.com/python/cpython/issues/50970#issuecomment-1147328694 seems to suggest that @vstinner has some appetite for addressing such issues and hence I'm filing this. **Your environment** - CPython versions tested on: 3.8.10, 3.10.4 - Operating system and architecture: Ubuntu 20.04, x86_64
process
multiprocessing with maxtasksperchild can hang if unpickling causes import bug report this seems like another specific instance of the general issue identified in if multiprocessing pool map async is used with maxtasksperchild and a value returned by a task is of a class not currently imported by the calling process it can lead to a hang here is an example that reliably hangs for me but which exits cleanly if elementtree is imported at the top level python usr bin env python import os import multiprocessing def worker num int from xml etree elementtree import elementtree print f worker num with pid os getpid return elementtree def main cores int num int pool multiprocessing pool processes cores maxtasksperchild barlist list pool map async worker list range num get print barlist if name main main running py spy dump on one of the workers shows this backtrace process python demo core py python usr bin thread idle thread handle workers acquire enter find and load worker demo core py mapstar multiprocessing pool py worker multiprocessing pool py run multiprocessing process py bootstrap multiprocessing process py launch multiprocessing popen fork py init multiprocessing popen fork py popen multiprocessing context py start multiprocessing process py repopulate pool static multiprocessing pool py maintain pool multiprocessing pool py handle workers multiprocessing pool py run threading py bootstrap inner threading py bootstrap threading py my guess without any further proof is that the main process receives a pickled elementtree and starts importing the module concurrently another thread realises it needs to start a new worker so does a fork the child process has a half imported locked elementtree module and tries to import it again leading to a deadlock note that this is nothing to do with elementtree i get the same behaviour with numpy i chose elementtree as a reasonably complex module to maximise the window for the race condition with a picklable class personally i consider the fork model of multiprocessing to be dangerous and requiring of care to ensure all worker tasks are created before doing anything that can conceivably create threads and definitely a bad combination with maxtasksperchild so i won t shed any tears if the resolution is won t fix don t do that but seems to suggest that vstinner has some appetite for addressing such issues and hence i m filing this your environment cpython versions tested on operating system and architecture ubuntu
1
21,598
30,002,906,066
IssuesEvent
2023-06-26 10:28:25
ukri-excalibur/excalibur-tests
https://api.github.com/repos/ukri-excalibur/excalibur-tests
opened
CI is failing after ReFrame 4.3.0 release
bug postprocessing
Our CI runs are failing to find the perflogs, maybe they have changed something in the naming convention. https://github.com/reframe-hpc/reframe/releases/tag/v4.3.0
1.0
CI is failing after ReFrame 4.3.0 release - Our CI runs are failing to find the perflogs, maybe they have changed something in the naming convention. https://github.com/reframe-hpc/reframe/releases/tag/v4.3.0
process
ci is failing after reframe release our ci runs are failing to find the perflogs maybe they have changed something in the naming convention
1
118,225
25,276,756,702
IssuesEvent
2022-11-16 13:11:14
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
reopened
Revisit the logic around couldShowWarning in InstallWarning
component: code quality priority: p3
Quoting @willdurand in https://github.com/mozilla/addons-frontend/pull/8348#discussion_r312547172, "I am not sure to like this pattern. Could you file an issue so that we can discuss it and maybe revisit it later?"
1.0
Revisit the logic around couldShowWarning in InstallWarning - Quoting @willdurand in https://github.com/mozilla/addons-frontend/pull/8348#discussion_r312547172, "I am not sure to like this pattern. Could you file an issue so that we can discuss it and maybe revisit it later?"
non_process
revisit the logic around couldshowwarning in installwarning quoting willdurand in i am not sure to like this pattern could you file an issue so that we can discuss it and maybe revisit it later
0
16,625
10,545,871,733
IssuesEvent
2019-10-02 20:08:05
cityofaustin/atd-geospatial
https://api.github.com/repos/cityofaustin/atd-geospatial
closed
Upgrade all computers to ArcGIS Desktop 10.6
Need: 1-Must Have Service: Geo Type: IT Support Workgroup: DTS
I will be working with Scott to upgrade all ATD computers from ArcGIS 10.x to 10.6. Before upgrading need to do the following steps: - Get a list of all users from Scott - Email all users notifying them of the upcoming upgrade, asking any users with custom tools to let me know so we can test before upgrading the user and having issues. - Work with any users that have custom tools to test out in 10.6 - Upgrade the DTS team first so we can ensure there are no issues with the new version before deploying to the entire department. - Notify users of saving a screenshot of ArcMap or creating a list of all toolbars they normally use and how they like their map document setup. Also what Add-Ins they have installed and use. - Work with Scott on getting all computers upgraded - Help users after the upgrade has taken place with any issues, reconnecting of databases, setup of ArcMap, Add-Ins, etc.
1.0
Upgrade all computers to ArcGIS Desktop 10.6 - I will be working with Scott to upgrade all ATD computers from ArcGIS 10.x to 10.6. Before upgrading need to do the following steps: - Get a list of all users from Scott - Email all users notifying them of the upcoming upgrade, asking any users with custom tools to let me know so we can test before upgrading the user and having issues. - Work with any users that have custom tools to test out in 10.6 - Upgrade the DTS team first so we can ensure there are no issues with the new version before deploying to the entire department. - Notify users of saving a screenshot of ArcMap or creating a list of all toolbars they normally use and how they like their map document setup. Also what Add-Ins they have installed and use. - Work with Scott on getting all computers upgraded - Help users after the upgrade has taken place with any issues, reconnecting of databases, setup of ArcMap, Add-Ins, etc.
non_process
upgrade all computers to arcgis desktop i will be working with scott to upgrade all atd computers from arcgis x to before upgrading need to do the following steps get a list of all users from scott email all users notifying them of the upcoming upgrade asking any users with custom tools to let me know so we can test before upgrading the user and having issues work with any users that have custom tools to test out in upgrade the dts team first so we can ensure there are no issues with the new version before deploying to the entire department notify users of saving a screenshot of arcmap or creating a list of all toolbars they normally use and how they like their map document setup also what add ins they have installed and use work with scott on getting all computers upgraded help users after the upgrade has taken place with any issues reconnecting of databases setup of arcmap add ins etc
0