Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
142,122
| 13,016,726,320
|
IssuesEvent
|
2020-07-26 08:22:14
|
anitab-org/anitab-org.github.io
|
https://api.github.com/repos/anitab-org/anitab-org.github.io
|
closed
|
Improvement Of README.md File.
|
Category: Documentation/Training First Timers Only
|
### Description
As a developer,
While setting up the remote upstream, documentation was unclear.
The README.md file describes the following.
"When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command -
git remote add upstream https://github.com/anitab-org/anitab-org.github.io."
### Expected Result
"When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command -
**cd anitab-org.github.io**
git remote add upstream https://github.com/anitab-org/anitab-org.github.io"
### Mocks

|
1.0
|
Improvement Of README.md File. - ### Description
As a developer,
While setting up the remote upstream, documentation was unclear.
The README.md file describes the following.
"When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command -
git remote add upstream https://github.com/anitab-org/anitab-org.github.io."
### Expected Result
"When a repository is cloned, it has a default remote named origin that points to your fork on GitHub, not the original repository it was forked from. To keep track of the original repository, you should add another remote named upstream. For this project, it can be done by running the following command -
**cd anitab-org.github.io**
git remote add upstream https://github.com/anitab-org/anitab-org.github.io"
### Mocks

|
non_process
|
improvement of readme md file description as a developer while setting up the remote upstream documentation was unclear the readme md file describes the following when a repository is cloned it has a default remote named origin that points to your fork on github not the original repository it was forked from to keep track of the original repository you should add another remote named upstream for this project it can be done by running the following command git remote add upstream expected result when a repository is cloned it has a default remote named origin that points to your fork on github not the original repository it was forked from to keep track of the original repository you should add another remote named upstream for this project it can be done by running the following command cd anitab org github io git remote add upstream mocks
| 0
|
13,376
| 15,837,949,036
|
IssuesEvent
|
2021-04-06 21:35:06
|
googleapis/python-pubsub
|
https://api.github.com/repos/googleapis/python-pubsub
|
closed
|
Unit tests must not use/expect credentials from the environment
|
api: pubsub type: process
|
```bash
$ env | grep GOOGLE && echo YES || echo NO
NO
$ git log -1
commit 469ebaa3c449c881089dfc657da5902c1d031803 (HEAD -> master, origin/master, origin/HEAD)
Author: Peter Lamut <plamut@users.noreply.github.com>
Date: Fri Apr 2 09:26:10 2021 +0200
chore: regenerate GAPIC layer with latest changes (#345)
$ nox -e unit-3.8
nox > Running session unit-3.8
nox > Creating virtual environment (virtualenv) using python3.8 in .nox/unit-3-8
nox > pip install asyncmock pytest-asyncio
nox > pip install mock pytest pytest-cov
nox > pip install -e .
nox > py.test --quiet --junitxml=unit_3.8_sponge_log.xml --cov=google/cloud --cov=tests/unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit
........................................................................ [ 8%]
........................................................................ [ 17%]
........................................................................ [ 26%]
........................................................................ [ 34%]
........................................................................ [ 43%]
........................................................................ [ 52%]
........................................................................ [ 61%]
........................................................................ [ 69%]
........................................................................ [ 78%]
........................................................................ [ 87%]
........................................................................ [ 96%]
............................... [100%]
=============================== warnings summary ===============================
...
tests/unit/pubsub_v1/publisher/test_publisher_client.py: 4 warnings
tests/unit/pubsub_v1/subscriber/test_subscriber_client.py: 9 warnings
/home/tseaver/projects/agendaless/Google/src/python-pubsub/.nox/unit-3-8/lib/python3.8/site-packages/google/auth/_default.py:70: UserWarning: Your application has authenticated using end user credentials from Google Cloud SDK without a quota project. You might receive a "quota exceeded" or "API not enabled" error. We recommend you rerun `gcloud auth application-default login` and make sure a quota project is added. Or you can use service accounts instead. For more information about service accounts, see https://cloud.google.com/docs/authentication/
warnings.warn(_CLOUD_SDK_CREDENTIALS_WARNING)
-- Docs: https://docs.pytest.org/en/stable/warnings.html
- generated xml file: /home/tseaver/projects/agendaless/Google/src/python-pubsub/unit_3.8_sponge_log.xml -
823 passed, 15 warnings in 13.94s
```
Unit tests should always pass explicit dummy credentials (e.g., see [`test_init`](https://github.com/googleapis/python-pubsub/blob/469ebaa3c449c881089dfc657da5902c1d031803/tests/unit/pubsub_v1/publisher/test_publisher_client.py#L54-L56)).
|
1.0
|
Unit tests must not use/expect credentials from the environment - ```bash
$ env | grep GOOGLE && echo YES || echo NO
NO
$ git log -1
commit 469ebaa3c449c881089dfc657da5902c1d031803 (HEAD -> master, origin/master, origin/HEAD)
Author: Peter Lamut <plamut@users.noreply.github.com>
Date: Fri Apr 2 09:26:10 2021 +0200
chore: regenerate GAPIC layer with latest changes (#345)
$ nox -e unit-3.8
nox > Running session unit-3.8
nox > Creating virtual environment (virtualenv) using python3.8 in .nox/unit-3-8
nox > pip install asyncmock pytest-asyncio
nox > pip install mock pytest pytest-cov
nox > pip install -e .
nox > py.test --quiet --junitxml=unit_3.8_sponge_log.xml --cov=google/cloud --cov=tests/unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit
........................................................................ [ 8%]
........................................................................ [ 17%]
........................................................................ [ 26%]
........................................................................ [ 34%]
........................................................................ [ 43%]
........................................................................ [ 52%]
........................................................................ [ 61%]
........................................................................ [ 69%]
........................................................................ [ 78%]
........................................................................ [ 87%]
........................................................................ [ 96%]
............................... [100%]
=============================== warnings summary ===============================
...
tests/unit/pubsub_v1/publisher/test_publisher_client.py: 4 warnings
tests/unit/pubsub_v1/subscriber/test_subscriber_client.py: 9 warnings
/home/tseaver/projects/agendaless/Google/src/python-pubsub/.nox/unit-3-8/lib/python3.8/site-packages/google/auth/_default.py:70: UserWarning: Your application has authenticated using end user credentials from Google Cloud SDK without a quota project. You might receive a "quota exceeded" or "API not enabled" error. We recommend you rerun `gcloud auth application-default login` and make sure a quota project is added. Or you can use service accounts instead. For more information about service accounts, see https://cloud.google.com/docs/authentication/
warnings.warn(_CLOUD_SDK_CREDENTIALS_WARNING)
-- Docs: https://docs.pytest.org/en/stable/warnings.html
- generated xml file: /home/tseaver/projects/agendaless/Google/src/python-pubsub/unit_3.8_sponge_log.xml -
823 passed, 15 warnings in 13.94s
```
Unit tests should always pass explicit dummy credentials (e.g., see [`test_init`](https://github.com/googleapis/python-pubsub/blob/469ebaa3c449c881089dfc657da5902c1d031803/tests/unit/pubsub_v1/publisher/test_publisher_client.py#L54-L56)).
|
process
|
unit tests must not use expect credentials from the environment bash env grep google echo yes echo no no git log commit head master origin master origin head author peter lamut date fri apr chore regenerate gapic layer with latest changes nox e unit nox running session unit nox creating virtual environment virtualenv using in nox unit nox pip install asyncmock pytest asyncio nox pip install mock pytest pytest cov nox pip install e nox py test quiet junitxml unit sponge log xml cov google cloud cov tests unit cov append cov config coveragerc cov report cov fail under tests unit warnings summary tests unit pubsub publisher test publisher client py warnings tests unit pubsub subscriber test subscriber client py warnings home tseaver projects agendaless google src python pubsub nox unit lib site packages google auth default py userwarning your application has authenticated using end user credentials from google cloud sdk without a quota project you might receive a quota exceeded or api not enabled error we recommend you rerun gcloud auth application default login and make sure a quota project is added or you can use service accounts instead for more information about service accounts see warnings warn cloud sdk credentials warning docs generated xml file home tseaver projects agendaless google src python pubsub unit sponge log xml passed warnings in unit tests should always pass explicit dummy credentials e g see
| 1
|
621,349
| 19,583,498,279
|
IssuesEvent
|
2022-01-05 01:49:08
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.donateblood.com.au - Design is broken
|
browser-firefox priority-normal severity-critical engine-gecko
|
<!-- @browser: Firefox 90.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:90.0) Gecko/20100101 Firefox/90.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/81122 -->
**URL**: https://www.donateblood.com.au/
**Browser / Version**: Firefox 90.0
**Operating System**: Mac OS X 10.15
**Tested Another Browser**: Yes Safari
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
I loaded the website by copying the URL https://www.donateblood.com.au/ into the URL bar in Firefox and hitting the ENTER key.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/7/2d88002f-203b-4de6-848c-dcdbaa651611.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.donateblood.com.au - Design is broken - <!-- @browser: Firefox 90.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:90.0) Gecko/20100101 Firefox/90.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/81122 -->
**URL**: https://www.donateblood.com.au/
**Browser / Version**: Firefox 90.0
**Operating System**: Mac OS X 10.15
**Tested Another Browser**: Yes Safari
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
I loaded the website by copying the URL https://www.donateblood.com.au/ into the URL bar in Firefox and hitting the ENTER key.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/7/2d88002f-203b-4de6-848c-dcdbaa651611.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
design is broken url browser version firefox operating system mac os x tested another browser yes safari problem type design is broken description items are overlapped steps to reproduce i loaded the website by copying the url into the url bar in firefox and hitting the enter key view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
9,718
| 12,716,601,222
|
IssuesEvent
|
2020-06-24 02:26:46
|
OUDcollective/twenty20times
|
https://api.github.com/repos/OUDcollective/twenty20times
|
opened
|
Understanding the GitHub flow · GitHub Guides
|
workflow-process
|

## Deploy
With GitHub, you can deploy from a branch for final testing in production before merging to master.
---
**Source URL**:
[https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
|
1.0
|
Understanding the GitHub flow · GitHub Guides - 
## Deploy
With GitHub, you can deploy from a branch for final testing in production before merging to master.
---
**Source URL**:
[https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
|
process
|
understanding the github flow · github guides deploy with github you can deploy from a branch for final testing in production before merging to master source url browser chrome os windows bit screen size viewport size pixel ratio zoom level
| 1
|
11,660
| 14,525,321,069
|
IssuesEvent
|
2020-12-14 12:46:53
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
closed
|
Add de-dot processor that converts dotted field names to nested objects
|
:Processors Filebeat Team:Services enhancement
|
**Background**
This is a requirement that came up in the https://github.com/elastic/ecs-logging initiative. In summary, we're trying to make logging simpler by logging ECS-compliant JSON to a file that Filebeat can just forward to Elasticsearch.
However, due to readability, performance, and other technical reasons, it's not always possible for the loggers to produce a correctly nested JSON structure. Some fields, like `log.logger` may be represented via a field name containing a dot (`"log.logger": "INFO"`) while others may be nested (`"foo": { "bar": "baz"}`).
More context here: https://github.com/elastic/ecs-logging-java/issues/51
**The problem**
When processing log data, for example with an Elasticsearch Ingest pipeline, we need all fields to be nested. Otherwise, the user doesn't know whether to access a field via `doc["foo.bar"]` or via `doc["foo"]["bar"]`. We don't want users to have knowledge about which fields are nested vs dotted as this is an implementation detail that can vary with different `ecs-logging` implementations and may even change for the same implementation.
Also, ECS defines that fields should be always nested.
**Describe the enhancement:**
We'd like to have a Filebeat processor that expands all dotted field names to nested objects. This would decouple the representation in the log file from how the documents are supposed to look once they hit the ingest node processing pipeline.
**Concerns**
- Performance: This might be a performance hit but I suspect other processors, like grok, to be much more processing intensive.
- If the JSON is already fully nested, we could short-circuit the processing
- We could require that once we're in a nested context, dots are no longer replaced with dotting.
- Allowed: `"foo.bar": {"baz": "qux"}`
- Disallowed: `"foo": {"bar.baz": "qux"}`
- Given this restricion, the de-dotting can be done very efficiently by sorting the keys alphabetically and processing the JSON similar to how a SAX-parser works (@urso's idea).
**Open Questions**
What to do when conflicts occur.
- Incompatible mappings: `"foo.bar": "baz"`, `"foo": "bar"` (foo is both an object and a string).
- Duplicate keys: `"foo.bar": "baz"`, `"foo.bar": "qux"`
|
1.0
|
Add de-dot processor that converts dotted field names to nested objects - **Background**
This is a requirement that came up in the https://github.com/elastic/ecs-logging initiative. In summary, we're trying to make logging simpler by logging ECS-compliant JSON to a file that Filebeat can just forward to Elasticsearch.
However, due to readability, performance, and other technical reasons, it's not always possible for the loggers to produce a correctly nested JSON structure. Some fields, like `log.logger` may be represented via a field name containing a dot (`"log.logger": "INFO"`) while others may be nested (`"foo": { "bar": "baz"}`).
More context here: https://github.com/elastic/ecs-logging-java/issues/51
**The problem**
When processing log data, for example with an Elasticsearch Ingest pipeline, we need all fields to be nested. Otherwise, the user doesn't know whether to access a field via `doc["foo.bar"]` or via `doc["foo"]["bar"]`. We don't want users to have knowledge about which fields are nested vs dotted as this is an implementation detail that can vary with different `ecs-logging` implementations and may even change for the same implementation.
Also, ECS defines that fields should be always nested.
**Describe the enhancement:**
We'd like to have a Filebeat processor that expands all dotted field names to nested objects. This would decouple the representation in the log file from how the documents are supposed to look once they hit the ingest node processing pipeline.
**Concerns**
- Performance: This might be a performance hit but I suspect other processors, like grok, to be much more processing intensive.
- If the JSON is already fully nested, we could short-circuit the processing
- We could require that once we're in a nested context, dots are no longer replaced with dotting.
- Allowed: `"foo.bar": {"baz": "qux"}`
- Disallowed: `"foo": {"bar.baz": "qux"}`
- Given this restricion, the de-dotting can be done very efficiently by sorting the keys alphabetically and processing the JSON similar to how a SAX-parser works (@urso's idea).
**Open Questions**
What to do when conflicts occur.
- Incompatible mappings: `"foo.bar": "baz"`, `"foo": "bar"` (foo is both an object and a string).
- Duplicate keys: `"foo.bar": "baz"`, `"foo.bar": "qux"`
|
process
|
add de dot processor that converts dotted field names to nested objects background this is a requirement that came up in the initiative in summary we re trying to make logging simpler by logging ecs compliant json to a file that filebeat can just forward to elasticsearch however due to readability performance and other technical reasons it s not always possible for the loggers to produce a correctly nested json structure some fields like log logger may be represented via a field name containing a dot log logger info while others may be nested foo bar baz more context here the problem when processing log data for example with an elasticsearch ingest pipeline we need all fields to be nested otherwise the user doesn t know whether to access a field via doc or via doc we don t want users to have knowledge about which fields are nested vs dotted as this is an implementation detail that can vary with different ecs logging implementations and may even change for the same implementation also ecs defines that fields should be always nested describe the enhancement we d like to have a filebeat processor that expands all dotted field names to nested objects this would decouple the representation in the log file from how the documents are supposed to look once they hit the ingest node processing pipeline concerns performance this might be a performance hit but i suspect other processors like grok to be much more processing intensive if the json is already fully nested we could short circuit the processing we could require that once we re in a nested context dots are no longer replaced with dotting allowed foo bar baz qux disallowed foo bar baz qux given this restricion the de dotting can be done very efficiently by sorting the keys alphabetically and processing the json similar to how a sax parser works urso s idea open questions what to do when conflicts occur incompatible mappings foo bar baz foo bar foo is both an object and a string duplicate keys foo bar baz foo bar qux
| 1
|
10,999
| 13,788,690,448
|
IssuesEvent
|
2020-10-09 07:40:11
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Uncaught TypeError: Illegal invocation
|
SYSTEM: iframe processing TYPE: bug support center
|
The page is not loaded, there is the following error in DevTools console:
```
Uncaught TypeError: Illegal invocation
at Window.addEventListener (hammerhead.js:7)
at <anonymous>:2355:89162
at Object.<anonymous> (<anonymous>:2355:89829)
at Object.jQuery (<anonymous>:2355:89881)
at n (<anonymous>:15:228)
at Object.<anonymous> (<anonymous>:2355:87883)
at Object.jQuery (<anonymous>:2355:88125)
at n (<anonymous>:15:228)
at Object.jQuery (<anonymous>:2355:87042)
at n (<anonymous>:15:228)
```
Please see the [T915177](https://supportcenter.devexpress.com/internal/ticket/details/T915177) private ticket for details.
testcafe-hammerhead version: 17.1.11
|
1.0
|
Uncaught TypeError: Illegal invocation - The page is not loaded, there is the following error in DevTools console:
```
Uncaught TypeError: Illegal invocation
at Window.addEventListener (hammerhead.js:7)
at <anonymous>:2355:89162
at Object.<anonymous> (<anonymous>:2355:89829)
at Object.jQuery (<anonymous>:2355:89881)
at n (<anonymous>:15:228)
at Object.<anonymous> (<anonymous>:2355:87883)
at Object.jQuery (<anonymous>:2355:88125)
at n (<anonymous>:15:228)
at Object.jQuery (<anonymous>:2355:87042)
at n (<anonymous>:15:228)
```
Please see the [T915177](https://supportcenter.devexpress.com/internal/ticket/details/T915177) private ticket for details.
testcafe-hammerhead version: 17.1.11
|
process
|
uncaught typeerror illegal invocation the page is not loaded there is the following error in devtools console uncaught typeerror illegal invocation at window addeventlistener hammerhead js at at object at object jquery at n at object at object jquery at n at object jquery at n please see the private ticket for details testcafe hammerhead version
| 1
|
214,966
| 24,126,376,888
|
IssuesEvent
|
2022-09-21 01:04:29
|
dmartinez777/Tracking
|
https://api.github.com/repos/dmartinez777/Tracking
|
closed
|
WS-2021-0013 (Medium) detected in laravel/framework-v5.8.35 - autoclosed
|
security vulnerability
|
## WS-2021-0013 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.8.35</b></p></summary>
<p>The Laravel Framework.</p>
<p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197">https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197</a></p>
<p>
Dependency Hierarchy:
- :x: **laravel/framework-v5.8.35** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Laravel is a web application framework. Versions of Laravel before 6.20.14, 7.30.4 and 8.24.0 contain a query binding exploitation.
If a request is crafted where a field that is normally a non-array value is an array, and that input is not validated or cast to its expected type before being passed to the query builder, an unexpected number of query bindings can be added to the query. In some situations, this will simply lead to no results being returned by the query builder; however, it is possible certain queries could be affected in a way that causes the query to return unexpected results.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://github.com/laravel/framework/commit/2d9b970257bca7a176be897ec18dd5f6ffc5497f>WS-2021-0013</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-x7p5-p2c9-phvg">https://github.com/advisories/GHSA-x7p5-p2c9-phvg</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: laravel/framework - 6.20.14, 7.30.4, 8.24.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0013 (Medium) detected in laravel/framework-v5.8.35 - autoclosed - ## WS-2021-0013 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.8.35</b></p></summary>
<p>The Laravel Framework.</p>
<p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197">https://api.github.com/repos/laravel/framework/zipball/5a9e4d241a8b815e16c9d2151e908992c38db197</a></p>
<p>
Dependency Hierarchy:
- :x: **laravel/framework-v5.8.35** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Laravel is a web application framework. Versions of Laravel before 6.20.14, 7.30.4 and 8.24.0 contain a query binding exploitation.
If a request is crafted where a field that is normally a non-array value is an array, and that input is not validated or cast to its expected type before being passed to the query builder, an unexpected number of query bindings can be added to the query. In some situations, this will simply lead to no results being returned by the query builder; however, it is possible certain queries could be affected in a way that causes the query to return unexpected results.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://github.com/laravel/framework/commit/2d9b970257bca7a176be897ec18dd5f6ffc5497f>WS-2021-0013</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-x7p5-p2c9-phvg">https://github.com/advisories/GHSA-x7p5-p2c9-phvg</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: laravel/framework - 6.20.14, 7.30.4, 8.24.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in laravel framework autoclosed ws medium severity vulnerability vulnerable library laravel framework the laravel framework library home page a href dependency hierarchy x laravel framework vulnerable library vulnerability details laravel is a web application framework versions of laravel before and contain a query binding exploitation if a request is crafted where a field that is normally a non array value is an array and that input is not validated or cast to its expected type before being passed to the query builder an unexpected number of query bindings can be added to the query in some situations this will simply lead to no results being returned by the query builder however it is possible certain queries could be affected in a way that causes the query to return unexpected results publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope changed impact metrics confidentiality impact high integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution laravel framework step up your open source security game with mend
| 0
|
19,566
| 25,887,826,902
|
IssuesEvent
|
2022-12-14 15:43:29
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_cuda_simple (__main__.TestMultiprocessing)
|
module: multiprocessing module: cuda triaged module: flaky-tests skipped
|
Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_cuda_simple&suite=TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8793519436).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_cuda_simple`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
cc @VitalyFedyunin @ngimel
|
1.0
|
DISABLED test_cuda_simple (__main__.TestMultiprocessing) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_cuda_simple&suite=TestMultiprocessing) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8793519436).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_cuda_simple`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
cc @VitalyFedyunin @ngimel
|
process
|
disabled test cuda simple main testmultiprocessing platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test cuda simple there should be several instances run as flaky tests are rerun in ci from which you can study the logs cc vitalyfedyunin ngimel
| 1
|
60,140
| 3,120,782,434
|
IssuesEvent
|
2015-09-05 01:59:04
|
framingeinstein/issues-test
|
https://api.github.com/repos/framingeinstein/issues-test
|
opened
|
SRP-17: Create Custom Admin Test Functions
|
priority:normal resolution:in-progress type:enhancement
|
Hi Meghan,
It looks like we can create some custom code to setup an admin-only payment method that will allow you to create test orders to send test emails. This would take up to an hour of billable work. If you want us to proceed, just let me know, and I'll get it in the que (but not for 1.1, at least as of now).
Let me know if you have any questions,
Thanks,
Abe
|
1.0
|
SRP-17: Create Custom Admin Test Functions - Hi Meghan,
It looks like we can create some custom code to setup an admin-only payment method that will allow you to create test orders to send test emails. This would take up to an hour of billable work. If you want us to proceed, just let me know, and I'll get it in the que (but not for 1.1, at least as of now).
Let me know if you have any questions,
Thanks,
Abe
|
non_process
|
srp create custom admin test functions hi meghan it looks like we can create some custom code to setup an admin only payment method that will allow you to create test orders to send test emails this would take up to an hour of billable work if you want us to proceed just let me know and i ll get it in the que but not for at least as of now let me know if you have any questions thanks abe
| 0
|
89,138
| 3,790,062,917
|
IssuesEvent
|
2016-03-21 20:08:05
|
ReactiveX/rxjs
|
https://api.github.com/repos/ReactiveX/rxjs
|
opened
|
Mono-repo, many packages
|
priority: critical
|
We want to convert the repository into a single repository from which we publish many packages. This means a reorganization of the repository to something more suitable.
Any and all ideas here are welcome. Especially those that come from prior art.
|
1.0
|
Mono-repo, many packages - We want to convert the repository into a single repository from which we publish many packages. This means a reorganization of the repository to something more suitable.
Any and all ideas here are welcome. Especially those that come from prior art.
|
non_process
|
mono repo many packages we want to convert the repository into a single repository from which we publish many packages this means a reorganization of the repository to something more suitable any and all ideas here are welcome especially those that come from prior art
| 0
|
445,094
| 12,826,084,727
|
IssuesEvent
|
2020-07-06 15:57:43
|
eclipse/dirigible
|
https://api.github.com/repos/eclipse/dirigible
|
closed
|
[EDM] Projection Entity type to be introduced
|
component-ide efforts-low priority-medium templates usability web-ide
|
An Entity which belongs to an external model (file) and it is used only as a projection.
No editing of any attribute or property is allowed in this case.
No generation of any output artefact is expected for this type.
|
1.0
|
[EDM] Projection Entity type to be introduced - An Entity which belongs to an external model (file) and it is used only as a projection.
No editing of any attribute or property is allowed in this case.
No generation of any output artefact is expected for this type.
|
non_process
|
projection entity type to be introduced an entity which belongs to an external model file and it is used only as a projection no editing of any attribute or property is allowed in this case no generation of any output artefact is expected for this type
| 0
|
107,278
| 23,382,330,095
|
IssuesEvent
|
2022-08-11 10:38:58
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Test failure JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh
|
arch-arm64 os-mac-os-x GCStress area-CodeGen-coreclr blocking-clean-ci-optional
|
Run: [runtime-coreclr gcstress-extra 20220724.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1900569&view=ms.vss-test-web.build-test-results-tab&runId=49457008&paneView=debug&resultId=108429)
Failed test:
```
coreclr OSX arm64 Checked gcstress0xc_zapdisable_heapverify1 @ OSX.1200.ARM64.Open
- JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh
```
**Error message:**
```
[createdump] Invalid process id: task_for_pid(4529) FAILED (os/kern) failure (5)
[createdump] This failure may be because createdump or the application is not properly signed and entitled.
[createdump] Failure took 0ms
/private/tmp/helix/working/A9B009AA/w/B9F709BA/e/JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh: line 373: 4529 Segmentation fault: 11 (core dumped) $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}"
Return code: 1
Raw output file: /tmp/helix/working/A9B009AA/w/B9F709BA/uploads/Reports/JIT.HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.output.txt
Raw output:
BEGIN EXECUTION
/tmp/helix/working/A9B009AA/p/corerun -p System.Reflection.Metadata.MetadataUpdater.IsSupported=false AdvSimd.Arm64_Part3_r.dll ''
Supported ISAs:
AdvSimd: True
Aes: True
ArmBase: True
Crc32: True
Dp: True
Rdm: True
Sha1: True
Sha256: True
Beginning test case MaxNumberPairwise.Vector128.Single at 7/24/2022 4:35:16 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScenario_Load
Beginning scenario: RunReflectionScenario_UnsafeRead
Beginning scenario: RunReflectionScenario_Load
Beginning scenario: RunClsVarScenario
Beginning scenario: RunClsVarScenario_Load
Beginning scenario: RunLclVarScenario_UnsafeRead
Beginning scenario: RunLclVarScenario_Load
Beginning scenario: RunClassLclFldScenario
Beginning scenario: RunClassLclFldScenario_Load
Beginning scenario: RunClassFldScenario
Beginning scenario: RunClassFldScenario_Load
Beginning scenario: RunStructLclFldScenario
Beginning scenario: RunStructLclFldScenario_Load
Beginning scenario: RunStructFldScenario
Beginning scenario: RunStructFldScenario_Load
Ending test case at 7/24/2022 4:35:26 PM
Beginning test case MaxNumberPairwiseScalar.Vector64.Single at 7/24/2022 4:35:26 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScenario_Load
Beginning scenario: RunReflectionScenario_UnsafeRead
Beginning scenario: RunReflectionScenario_Load
Beginning scenario: RunClsVarScenario
Beginning scenario: RunClsVarScenario_Load
Beginning scenario: RunLclVarScenario_UnsafeRead
Beginning scenario: RunLclVarScenario_Load
Beginning scenario: RunClassLclFldScenario
Beginning scenario: RunClassLclFldScenario_Load
Beginning scenario: RunClassFldScenario
Beginning scenario: RunClassFldScenario_Load
Beginning scenario: RunStructLclFldScenario
Beginning scenario: RunStructLclFldScenario_Load
Beginning scenario: RunStructFldScenario
Beginning scenario: RunStructFldScenario_Load
Ending test case at 7/24/2022 4:35:28 PM
Beginning test case MaxNumberPairwiseScalar.Vector128.Double at 7/24/2022 4:35:28 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScenario_Load
Beginning scenario: RunReflectionScenario_UnsafeRead
Beginning scenario: RunReflectionScenario_Load
Beginning scenario: RunClsVarScenario
Beginning scenario: RunClsVarScenario_Load
Beginning scenario: RunLclVarScenario_UnsafeRead
Beginning scenario: RunLclVarScenario_Load
Beginning scenario: RunClassLclFldScenario
Beginning scenario: RunClassLclFldScenario_Load
Beginning scenario: RunClassFldScenario
Beginning scenario: RunClassFldScenario_Load
Beginning scenario: RunStructLclFldScenario
Beginning scenario: RunStructLclFldScenario_Load
Beginning scenario: RunStructFldScenario
Beginning scenario: RunStructFldScenario_Load
Ending test case at 7/24/2022 4:35:30 PM
Beginning test case MaxPairwise.Vector128.Byte at 7/24/2022 4:35:30 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScen
Stack trace
at JIT_HardwareIntrinsics._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_sh()
at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr)
```
|
1.0
|
Test failure JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh - Run: [runtime-coreclr gcstress-extra 20220724.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1900569&view=ms.vss-test-web.build-test-results-tab&runId=49457008&paneView=debug&resultId=108429)
Failed test:
```
coreclr OSX arm64 Checked gcstress0xc_zapdisable_heapverify1 @ OSX.1200.ARM64.Open
- JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh
```
**Error message:**
```
[createdump] Invalid process id: task_for_pid(4529) FAILED (os/kern) failure (5)
[createdump] This failure may be because createdump or the application is not properly signed and entitled.
[createdump] Failure took 0ms
/private/tmp/helix/working/A9B009AA/w/B9F709BA/e/JIT/HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.sh: line 373: 4529 Segmentation fault: 11 (core dumped) $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}"
Return code: 1
Raw output file: /tmp/helix/working/A9B009AA/w/B9F709BA/uploads/Reports/JIT.HardwareIntrinsics/Arm/AdvSimd.Arm64/AdvSimd.Arm64_Part3_r/AdvSimd.Arm64_Part3_r.output.txt
Raw output:
BEGIN EXECUTION
/tmp/helix/working/A9B009AA/p/corerun -p System.Reflection.Metadata.MetadataUpdater.IsSupported=false AdvSimd.Arm64_Part3_r.dll ''
Supported ISAs:
AdvSimd: True
Aes: True
ArmBase: True
Crc32: True
Dp: True
Rdm: True
Sha1: True
Sha256: True
Beginning test case MaxNumberPairwise.Vector128.Single at 7/24/2022 4:35:16 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScenario_Load
Beginning scenario: RunReflectionScenario_UnsafeRead
Beginning scenario: RunReflectionScenario_Load
Beginning scenario: RunClsVarScenario
Beginning scenario: RunClsVarScenario_Load
Beginning scenario: RunLclVarScenario_UnsafeRead
Beginning scenario: RunLclVarScenario_Load
Beginning scenario: RunClassLclFldScenario
Beginning scenario: RunClassLclFldScenario_Load
Beginning scenario: RunClassFldScenario
Beginning scenario: RunClassFldScenario_Load
Beginning scenario: RunStructLclFldScenario
Beginning scenario: RunStructLclFldScenario_Load
Beginning scenario: RunStructFldScenario
Beginning scenario: RunStructFldScenario_Load
Ending test case at 7/24/2022 4:35:26 PM
Beginning test case MaxNumberPairwiseScalar.Vector64.Single at 7/24/2022 4:35:26 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScenario_Load
Beginning scenario: RunReflectionScenario_UnsafeRead
Beginning scenario: RunReflectionScenario_Load
Beginning scenario: RunClsVarScenario
Beginning scenario: RunClsVarScenario_Load
Beginning scenario: RunLclVarScenario_UnsafeRead
Beginning scenario: RunLclVarScenario_Load
Beginning scenario: RunClassLclFldScenario
Beginning scenario: RunClassLclFldScenario_Load
Beginning scenario: RunClassFldScenario
Beginning scenario: RunClassFldScenario_Load
Beginning scenario: RunStructLclFldScenario
Beginning scenario: RunStructLclFldScenario_Load
Beginning scenario: RunStructFldScenario
Beginning scenario: RunStructFldScenario_Load
Ending test case at 7/24/2022 4:35:28 PM
Beginning test case MaxNumberPairwiseScalar.Vector128.Double at 7/24/2022 4:35:28 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScenario_Load
Beginning scenario: RunReflectionScenario_UnsafeRead
Beginning scenario: RunReflectionScenario_Load
Beginning scenario: RunClsVarScenario
Beginning scenario: RunClsVarScenario_Load
Beginning scenario: RunLclVarScenario_UnsafeRead
Beginning scenario: RunLclVarScenario_Load
Beginning scenario: RunClassLclFldScenario
Beginning scenario: RunClassLclFldScenario_Load
Beginning scenario: RunClassFldScenario
Beginning scenario: RunClassFldScenario_Load
Beginning scenario: RunStructLclFldScenario
Beginning scenario: RunStructLclFldScenario_Load
Beginning scenario: RunStructFldScenario
Beginning scenario: RunStructFldScenario_Load
Ending test case at 7/24/2022 4:35:30 PM
Beginning test case MaxPairwise.Vector128.Byte at 7/24/2022 4:35:30 PM
Random seed: 20010415; set environment variable CORECLR_SEED to this value to repro
Beginning scenario: RunBasicScenario_UnsafeRead
Beginning scenario: RunBasicScen
Stack trace
at JIT_HardwareIntrinsics._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_._Arm_AdvSimd_Arm64_AdvSimd_Arm64_Part3_r_AdvSimd_Arm64_Part3_r_sh()
at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr)
```
|
non_process
|
test failure jit hardwareintrinsics arm advsimd advsimd r advsimd r sh run failed test coreclr osx checked zapdisable osx open jit hardwareintrinsics arm advsimd advsimd r advsimd r sh error message invalid process id task for pid failed os kern failure this failure may be because createdump or the application is not properly signed and entitled failure took private tmp helix working w e jit hardwareintrinsics arm advsimd advsimd r advsimd r sh line segmentation fault core dumped launcher exepath clrtestexecutionarguments return code raw output file tmp helix working w uploads reports jit hardwareintrinsics arm advsimd advsimd r advsimd r output txt raw output begin execution tmp helix working p corerun p system reflection metadata metadataupdater issupported false advsimd r dll supported isas advsimd true aes true armbase true true dp true rdm true true true beginning test case maxnumberpairwise single at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscenario load beginning scenario runreflectionscenario unsaferead beginning scenario runreflectionscenario load beginning scenario runclsvarscenario beginning scenario runclsvarscenario load beginning scenario runlclvarscenario unsaferead beginning scenario runlclvarscenario load beginning scenario runclasslclfldscenario beginning scenario runclasslclfldscenario load beginning scenario runclassfldscenario beginning scenario runclassfldscenario load beginning scenario runstructlclfldscenario beginning scenario runstructlclfldscenario load beginning scenario runstructfldscenario beginning scenario runstructfldscenario load ending test case at pm beginning test case maxnumberpairwisescalar single at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscenario load beginning scenario runreflectionscenario unsaferead beginning scenario runreflectionscenario load beginning scenario runclsvarscenario beginning scenario runclsvarscenario load beginning scenario runlclvarscenario unsaferead beginning scenario runlclvarscenario load beginning scenario runclasslclfldscenario beginning scenario runclasslclfldscenario load beginning scenario runclassfldscenario beginning scenario runclassfldscenario load beginning scenario runstructlclfldscenario beginning scenario runstructlclfldscenario load beginning scenario runstructfldscenario beginning scenario runstructfldscenario load ending test case at pm beginning test case maxnumberpairwisescalar double at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscenario load beginning scenario runreflectionscenario unsaferead beginning scenario runreflectionscenario load beginning scenario runclsvarscenario beginning scenario runclsvarscenario load beginning scenario runlclvarscenario unsaferead beginning scenario runlclvarscenario load beginning scenario runclasslclfldscenario beginning scenario runclasslclfldscenario load beginning scenario runclassfldscenario beginning scenario runclassfldscenario load beginning scenario runstructlclfldscenario beginning scenario runstructlclfldscenario load beginning scenario runstructfldscenario beginning scenario runstructfldscenario load ending test case at pm beginning test case maxpairwise byte at pm random seed set environment variable coreclr seed to this value to repro beginning scenario runbasicscenario unsaferead beginning scenario runbasicscen stack trace at jit hardwareintrinsics arm advsimd advsimd r advsimd r arm advsimd advsimd r advsimd r sh at system runtimemethodhandle invokemethod object target void arguments signature sig boolean isconstructor at system reflection methodinvoker invoke object obj intptr args bindingflags invokeattr
| 0
|
11,757
| 14,591,552,135
|
IssuesEvent
|
2020-12-19 13:34:16
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
symfony/process returns empty outputs on IIS
|
Bug Process Status: Needs Review
|
| Q | A
| ---------------- | -----
| Bug report? | yes
| Feature request? | no
| BC Break report? | no
| RFC? | no
| Symfony version | 3.3.11
Hello,
I'm using PHP 7.1 on IIS 7.5 and symfony/process always returns an empty output.
Here is a test case:
```php
<?php
use Symfony\Component\Process\Process;
require_once __DIR__.'/vendor/autoload.php';
//This returns an output
var_dump(shell_exec('dir'));
$process = new Process('dir');
$process->mustRun();
//These return empty strings
var_dump(
$process->getOutput(),
$process->getErrorOutput()
);
```
What's strange is that it works when calling PHP (same binary) from the commandline.
And of course the same code works fine on my Linux/Apache server.
I tested various other commands (`cd`, `Python.exe`) and I always get the same empty result.
`mustRun()` does throw an exception if the command does not exist:
```
PHP Fatal error: Uncaught Symfony\Component\Process\Exception\ProcessFailedException: The command "foobar" failed.
Exit Code: 1(General error)
Working directory: C:\inetpub\wwwroot
Output:
================
Error Output:
================
in C:\inetpub\wwwroot\vendor\symfony\process\Process.php:241
Stack trace:
#0 C:\inetpub\wwwroot\test.php(11): Symfony\Component\Process\Process->mustRun()
#1 {main}
thrown in C:\inetpub\wwwroot\vendor\symfony\process\Process.php on line 241
```
Edit: I also tried configuring IIS to use another PHP binary (installed with Chocolatey) and I get the same issue, so I guess it is linked to IIS itself.
|
1.0
|
symfony/process returns empty outputs on IIS - | Q | A
| ---------------- | -----
| Bug report? | yes
| Feature request? | no
| BC Break report? | no
| RFC? | no
| Symfony version | 3.3.11
Hello,
I'm using PHP 7.1 on IIS 7.5 and symfony/process always returns an empty output.
Here is a test case:
```php
<?php
use Symfony\Component\Process\Process;
require_once __DIR__.'/vendor/autoload.php';
//This returns an output
var_dump(shell_exec('dir'));
$process = new Process('dir');
$process->mustRun();
//These return empty strings
var_dump(
$process->getOutput(),
$process->getErrorOutput()
);
```
What's strange is that it works when calling PHP (same binary) from the commandline.
And of course the same code works fine on my Linux/Apache server.
I tested various other commands (`cd`, `Python.exe`) and I always get the same empty result.
`mustRun()` does throw an exception if the command does not exist:
```
PHP Fatal error: Uncaught Symfony\Component\Process\Exception\ProcessFailedException: The command "foobar" failed.
Exit Code: 1(General error)
Working directory: C:\inetpub\wwwroot
Output:
================
Error Output:
================
in C:\inetpub\wwwroot\vendor\symfony\process\Process.php:241
Stack trace:
#0 C:\inetpub\wwwroot\test.php(11): Symfony\Component\Process\Process->mustRun()
#1 {main}
thrown in C:\inetpub\wwwroot\vendor\symfony\process\Process.php on line 241
```
Edit: I also tried configuring IIS to use another PHP binary (installed with Chocolatey) and I get the same issue, so I guess it is linked to IIS itself.
|
process
|
symfony process returns empty outputs on iis q a bug report yes feature request no bc break report no rfc no symfony version hello i m using php on iis and symfony process always returns an empty output here is a test case php php use symfony component process process require once dir vendor autoload php this returns an output var dump shell exec dir process new process dir process mustrun these return empty strings var dump process getoutput process geterroroutput what s strange is that it works when calling php same binary from the commandline and of course the same code works fine on my linux apache server i tested various other commands cd python exe and i always get the same empty result mustrun does throw an exception if the command does not exist php fatal error uncaught symfony component process exception processfailedexception the command foobar failed exit code general error working directory c inetpub wwwroot output error output in c inetpub wwwroot vendor symfony process process php stack trace c inetpub wwwroot test php symfony component process process mustrun main thrown in c inetpub wwwroot vendor symfony process process php on line edit i also tried configuring iis to use another php binary installed with chocolatey and i get the same issue so i guess it is linked to iis itself
| 1
|
92,620
| 8,373,219,296
|
IssuesEvent
|
2018-10-05 09:39:34
|
andrewwood2/acebook-gazelle
|
https://api.github.com/repos/andrewwood2/acebook-gazelle
|
closed
|
B. Set up front end feature testing framework
|
in progress test
|
Set up feature testing for React front end. either Mocha, Zombie, etc.
|
1.0
|
B. Set up front end feature testing framework - Set up feature testing for React front end. either Mocha, Zombie, etc.
|
non_process
|
b set up front end feature testing framework set up feature testing for react front end either mocha zombie etc
| 0
|
41,197
| 10,331,326,860
|
IssuesEvent
|
2019-09-02 17:33:01
|
davidjamesca/ctypesgen
|
https://api.github.com/repos/davidjamesca/ctypesgen
|
closed
|
Tests failing to find libc.so.6 and libm.so.6 on 64bit Ubuntu (SVN r147)
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. Checkout the code
2. Go into the test directory
3. Run "./testsuite.py"
What is the expected output? What do you see instead?
I expect to see all tests pass instead I see 11 errors all like this:
======================================================================
ERROR: test_bad_args_string_not_number (__main__.MathTest)
Based on math_functions.py
----------------------------------------------------------------------
Traceback (most recent call last):
File "./testsuite.py", line 252, in setUp
self.module, output = ctypesgentest.test(header_str, libraries=libraries, all_headers=True)
File "/home/jlisee/projects/ctypesgen-read-only/test/ctypesgentest.py", line 52, in test
module = __import__("temp")
File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 598, in <module>
_libs["libm.so.6"] = load_library("libm.so.6")
File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 367, in load_library
raise ImportError("%s not found." % libname)
ImportError: libm.so.6 not found
What version of the product are you using? On what operating system?
Ubuntu 12.04 64bit, SVN r147.
Please provide any additional information below.
I have attached a patch to fix the issue. There are still 3 tests failing with
"AttributeError: type object 'c_uint' has no attribute '_fields_'" in the
generated "temp.py" file.
```
Original issue reported on code.google.com by `jli...@gmail.com` on 28 Feb 2013 at 3:02
Attachments:
- [Ubuntu_64bit_fix.patch](https://storage.googleapis.com/google-code-attachments/ctypesgen/issue-39/comment-0/Ubuntu_64bit_fix.patch)
|
1.0
|
Tests failing to find libc.so.6 and libm.so.6 on 64bit Ubuntu (SVN r147) - ```
What steps will reproduce the problem?
1. Checkout the code
2. Go into the test directory
3. Run "./testsuite.py"
What is the expected output? What do you see instead?
I expect to see all tests pass instead I see 11 errors all like this:
======================================================================
ERROR: test_bad_args_string_not_number (__main__.MathTest)
Based on math_functions.py
----------------------------------------------------------------------
Traceback (most recent call last):
File "./testsuite.py", line 252, in setUp
self.module, output = ctypesgentest.test(header_str, libraries=libraries, all_headers=True)
File "/home/jlisee/projects/ctypesgen-read-only/test/ctypesgentest.py", line 52, in test
module = __import__("temp")
File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 598, in <module>
_libs["libm.so.6"] = load_library("libm.so.6")
File "/home/jlisee/projects/ctypesgen-read-only/test/temp.py", line 367, in load_library
raise ImportError("%s not found." % libname)
ImportError: libm.so.6 not found
What version of the product are you using? On what operating system?
Ubuntu 12.04 64bit, SVN r147.
Please provide any additional information below.
I have attached a patch to fix the issue. There are still 3 tests failing with
"AttributeError: type object 'c_uint' has no attribute '_fields_'" in the
generated "temp.py" file.
```
Original issue reported on code.google.com by `jli...@gmail.com` on 28 Feb 2013 at 3:02
Attachments:
- [Ubuntu_64bit_fix.patch](https://storage.googleapis.com/google-code-attachments/ctypesgen/issue-39/comment-0/Ubuntu_64bit_fix.patch)
|
non_process
|
tests failing to find libc so and libm so on ubuntu svn what steps will reproduce the problem checkout the code go into the test directory run testsuite py what is the expected output what do you see instead i expect to see all tests pass instead i see errors all like this error test bad args string not number main mathtest based on math functions py traceback most recent call last file testsuite py line in setup self module output ctypesgentest test header str libraries libraries all headers true file home jlisee projects ctypesgen read only test ctypesgentest py line in test module import temp file home jlisee projects ctypesgen read only test temp py line in libs load library libm so file home jlisee projects ctypesgen read only test temp py line in load library raise importerror s not found libname importerror libm so not found what version of the product are you using on what operating system ubuntu svn please provide any additional information below i have attached a patch to fix the issue there are still tests failing with attributeerror type object c uint has no attribute fields in the generated temp py file original issue reported on code google com by jli gmail com on feb at attachments
| 0
|
757,685
| 26,524,390,024
|
IssuesEvent
|
2023-01-19 07:18:54
|
pystardust/ani-cli
|
https://api.github.com/repos/pystardust/ani-cli
|
closed
|
Episodes not released yet!
|
type: bug priority 2: medium
|
Version: 3.4.0
OS: Debian 11
Shell: Bash 5.1.4
Anime: JoJo
I get an error "Episodes not released yet!" no matter the anime I choose
1. Run `ani-cli JoJo
2. Choose 1
3. Episodes not released yet!
|
1.0
|
Episodes not released yet! - Version: 3.4.0
OS: Debian 11
Shell: Bash 5.1.4
Anime: JoJo
I get an error "Episodes not released yet!" no matter the anime I choose
1. Run `ani-cli JoJo
2. Choose 1
3. Episodes not released yet!
|
non_process
|
episodes not released yet version os debian shell bash anime jojo i get an error episodes not released yet no matter the anime i choose run ani cli jojo choose episodes not released yet
| 0
|
276,637
| 20,993,384,598
|
IssuesEvent
|
2022-03-29 11:25:21
|
Sitecore/developer-portal
|
https://api.github.com/repos/Sitecore/developer-portal
|
closed
|
Create DevOps guide for Managed Cloud Containers
|
documentation
|
Need an article added to the Getting Started section (`/learn/getting-started`) area that can pull together all the documentation and steps a developer may need to get started with container GitOps for Managed Cloud Containers.
We also don't have a Managed Cloud product area on the site yet, so that will need to be added as well so that this guide can be found.
|
1.0
|
Create DevOps guide for Managed Cloud Containers - Need an article added to the Getting Started section (`/learn/getting-started`) area that can pull together all the documentation and steps a developer may need to get started with container GitOps for Managed Cloud Containers.
We also don't have a Managed Cloud product area on the site yet, so that will need to be added as well so that this guide can be found.
|
non_process
|
create devops guide for managed cloud containers need an article added to the getting started section learn getting started area that can pull together all the documentation and steps a developer may need to get started with container gitops for managed cloud containers we also don t have a managed cloud product area on the site yet so that will need to be added as well so that this guide can be found
| 0
|
10,920
| 13,697,018,917
|
IssuesEvent
|
2020-10-01 01:46:37
|
opendistro-for-elasticsearch/opendistro-build
|
https://api.github.com/repos/opendistro-for-elasticsearch/opendistro-build
|
closed
|
Release ODFE 1.10.1 based on ES 7.9.1
|
in process infra new release
|
ODFE 1.10.1 based on ES 7.9.1
**(Note: ODFE 1.10.0 is skipped right now as we prefer to release ODFE 1.10.1 for ES 7.9.1. This is to avoid memory leak in Lucene 8.6.0 and 8.6.1 (ES 7.9.1 has Lucene 8.6.2) https://github.com/elastic/elasticsearch/issues/61512)**
Release Engineering / Build Repo Key Changes:
* KNNLib will now use wildcard to resolve hardcoded version issues ([#359](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/359))
* Docker allows elasticsearch user to access logs under supervisord folder ([#271](https://github.com/opendistro-for-elasticsearch/performance-analyzer-rca/pull/271), [#146](https://github.com/opendistro-for-elasticsearch/performance-analyzer/pull/146), [#320](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/320))
* Implement Version Cuts for consistent distribution release builds ([#357](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/357))
* Add descriptions for several scripts with usage documentations ([#334](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/334))
* Update opendistro-build github repo issues link ([#382](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/382))
* Disable optimizations for KNNLib compilation in docker image creation ([#384](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/384))
* HELM allows customizing docker registry, thanks @tareqhs ([#358](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/358))
* HELM Kibana ingress path fix, thanks @Hokwang ([#340](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/340))
* Helm master nodes allows extraVolumeMounts when securityconfig disabled, thanks @aplhk ([#366](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/366))
* HELM Readme Update, thanks @dmpe ([#380](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/380) [#385](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/385))
* Kibana has new cookie settings for security kibana plugin 2.0 framework ([#397](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/397))
------
### ODFE 1.10.1 Released as of 2020/09/30
* Downloads [Here](https://opendistro.github.io/for-elasticsearch/downloads.html)
* [1.10.1 Release Notes](https://github.com/opendistro-for-elasticsearch/opendistro-build/blob/master/release-notes/opendistro-for-elasticsearch-release-notes-1.10.1.md)
* [1.10.1 Blog Post](https://opendistro.github.io/for-elasticsearch/blog/odfe-updates/2020/09/Open-Distro-for-Elasticsearch-1.10.1-is-released/)
------
ODFE 1.10.0 post for backup purposes: https://github.com/opendistro-for-elasticsearch/opendistro-build/issues/350
|
1.0
|
Release ODFE 1.10.1 based on ES 7.9.1 - ODFE 1.10.1 based on ES 7.9.1
**(Note: ODFE 1.10.0 is skipped right now as we prefer to release ODFE 1.10.1 for ES 7.9.1. This is to avoid memory leak in Lucene 8.6.0 and 8.6.1 (ES 7.9.1 has Lucene 8.6.2) https://github.com/elastic/elasticsearch/issues/61512)**
Release Engineering / Build Repo Key Changes:
* KNNLib will now use wildcard to resolve hardcoded version issues ([#359](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/359))
* Docker allows elasticsearch user to access logs under supervisord folder ([#271](https://github.com/opendistro-for-elasticsearch/performance-analyzer-rca/pull/271), [#146](https://github.com/opendistro-for-elasticsearch/performance-analyzer/pull/146), [#320](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/320))
* Implement Version Cuts for consistent distribution release builds ([#357](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/357))
* Add descriptions for several scripts with usage documentations ([#334](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/334))
* Update opendistro-build github repo issues link ([#382](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/382))
* Disable optimizations for KNNLib compilation in docker image creation ([#384](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/384))
* HELM allows customizing docker registry, thanks @tareqhs ([#358](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/358))
* HELM Kibana ingress path fix, thanks @Hokwang ([#340](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/340))
* Helm master nodes allows extraVolumeMounts when securityconfig disabled, thanks @aplhk ([#366](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/366))
* HELM Readme Update, thanks @dmpe ([#380](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/380) [#385](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/385))
* Kibana has new cookie settings for security kibana plugin 2.0 framework ([#397](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/397))
------
### ODFE 1.10.1 Released as of 2020/09/30
* Downloads [Here](https://opendistro.github.io/for-elasticsearch/downloads.html)
* [1.10.1 Release Notes](https://github.com/opendistro-for-elasticsearch/opendistro-build/blob/master/release-notes/opendistro-for-elasticsearch-release-notes-1.10.1.md)
* [1.10.1 Blog Post](https://opendistro.github.io/for-elasticsearch/blog/odfe-updates/2020/09/Open-Distro-for-Elasticsearch-1.10.1-is-released/)
------
ODFE 1.10.0 post for backup purposes: https://github.com/opendistro-for-elasticsearch/opendistro-build/issues/350
|
process
|
release odfe based on es odfe based on es note odfe is skipped right now as we prefer to release odfe for es this is to avoid memory leak in lucene and es has lucene release engineering build repo key changes knnlib will now use wildcard to resolve hardcoded version issues docker allows elasticsearch user to access logs under supervisord folder implement version cuts for consistent distribution release builds add descriptions for several scripts with usage documentations update opendistro build github repo issues link disable optimizations for knnlib compilation in docker image creation helm allows customizing docker registry thanks tareqhs helm kibana ingress path fix thanks hokwang helm master nodes allows extravolumemounts when securityconfig disabled thanks aplhk helm readme update thanks dmpe kibana has new cookie settings for security kibana plugin framework odfe released as of downloads odfe post for backup purposes
| 1
|
69,037
| 7,122,685,713
|
IssuesEvent
|
2018-01-19 12:49:36
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
Flaky http2-settings-flood
|
CI / flaky test freebsd
|
Multiple timeouts on this test:
https://ci.nodejs.org/job/node-test-commit-freebsd/14756/nodes=freebsd10-64/console
https://ci.nodejs.org/job/node-test-commit-freebsd/14746/nodes=freebsd10-64/console
```
not ok 2035 sequential/test-http2-settings-flood
---
duration_ms: 120.105
severity: fail
stack: |-
timeout
```
|
1.0
|
Flaky http2-settings-flood - Multiple timeouts on this test:
https://ci.nodejs.org/job/node-test-commit-freebsd/14756/nodes=freebsd10-64/console
https://ci.nodejs.org/job/node-test-commit-freebsd/14746/nodes=freebsd10-64/console
```
not ok 2035 sequential/test-http2-settings-flood
---
duration_ms: 120.105
severity: fail
stack: |-
timeout
```
|
non_process
|
flaky settings flood multiple timeouts on this test not ok sequential test settings flood duration ms severity fail stack timeout
| 0
|
618,568
| 19,474,854,104
|
IssuesEvent
|
2021-12-24 10:07:40
|
MartinXPN/profound.academy
|
https://api.github.com/repos/MartinXPN/profound.academy
|
closed
|
Allow visitors to view exercises
|
enhancement priority/top
|
Allow visitors to view exercises but not submit them until they sign up.
In case they submit anything add the course to curriculum
|
1.0
|
Allow visitors to view exercises - Allow visitors to view exercises but not submit them until they sign up.
In case they submit anything add the course to curriculum
|
non_process
|
allow visitors to view exercises allow visitors to view exercises but not submit them until they sign up in case they submit anything add the course to curriculum
| 0
|
19,563
| 25,885,019,166
|
IssuesEvent
|
2022-12-14 13:59:51
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/spanmetrics] getting key from cache error
|
bug processor/spanmetrics
|
### What happened?
## Description
I get an error in my app when it sent span to a otel collector which configed a spanmetrics processor with `defaultDimensionsCacheSize`:
```
2022/10/27 02:37:46 rpc error: code = Unknown desc = value not found in metricKeyToDimensions cache by key "amamba\x00amamba.io.api.pipeline.v1alpha1.Pipelines/ReplayPipelineRun\x00SPAN_KIND_SERVER\x00STATUS_CODE_OK\x00cd7b102e-fbc5-4556-a0fd-718298df3de9\x00amamba-system\x00demo-dev-worker-03\x00amamba-apiserver-66c9486b55-5l4n8"
```
I trick the error to:
https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/92ad54f707a967386b764b6fdaea5a2ba4377319/processor/spanmetricsprocessor/processor.go#L335-L344
### Collector version
v0.61.0
### Environment information
## Environment
OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")
### OpenTelemetry Collector configuration
_No response_
### Log output
_No response_
### Additional context
resolve by #15687
|
1.0
|
[processor/spanmetrics] getting key from cache error - ### What happened?
## Description
I get an error in my app when it sent span to a otel collector which configed a spanmetrics processor with `defaultDimensionsCacheSize`:
```
2022/10/27 02:37:46 rpc error: code = Unknown desc = value not found in metricKeyToDimensions cache by key "amamba\x00amamba.io.api.pipeline.v1alpha1.Pipelines/ReplayPipelineRun\x00SPAN_KIND_SERVER\x00STATUS_CODE_OK\x00cd7b102e-fbc5-4556-a0fd-718298df3de9\x00amamba-system\x00demo-dev-worker-03\x00amamba-apiserver-66c9486b55-5l4n8"
```
I trick the error to:
https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/92ad54f707a967386b764b6fdaea5a2ba4377319/processor/spanmetricsprocessor/processor.go#L335-L344
### Collector version
v0.61.0
### Environment information
## Environment
OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")
### OpenTelemetry Collector configuration
_No response_
### Log output
_No response_
### Additional context
resolve by #15687
|
process
|
getting key from cache error what happened description i get an error in my app when it sent span to a otel collector which configed a spanmetrics processor with defaultdimensionscachesize rpc error code unknown desc value not found in metrickeytodimensions cache by key amamba io api pipeline pipelines replaypipelinerun kind server code ok system dev worker apiserver i trick the error to collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration no response log output no response additional context resolve by
| 1
|
325,176
| 9,920,360,218
|
IssuesEvent
|
2019-06-30 08:30:45
|
googleapis/elixir-google-api
|
https://api.github.com/repos/googleapis/elixir-google-api
|
closed
|
Synthesis failed for DLP
|
:rotating_light: autosynth failure priority: p1 triage me type: bug
|
Hello! Autosynth couldn't regenerate DLP. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth-dlp'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/dlp/synth.metadata', 'synth.py', '--']
synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py.
synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
synthtool > Failed executing git clean -fdx:
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/reviews.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/edits.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/orders.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/purchases.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/inappproducts.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/internalappsharingartifacts.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/connection.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundles_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_file.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundle.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/image.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk_binary.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listing.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/page_info.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/testers.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/season.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_details.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review_reply_result.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/tracks_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchase.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchase.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchases_acknowledge_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_cancel_survey_result.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/timestamp.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_files_upload_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_upload_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_price_change.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_deferral_info.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_acknowledge_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchase.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_delete_all_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/price.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/prorate.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk_uses_permission.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/developer_comment.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchases_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_file.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/token_pagination.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/localized_text.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/country_targeting.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/month_day.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_files_upload_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track_release.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/device_metadata.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product_listing.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_edit.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/internal_app_sharing_artifact.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/inappproducts_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/user_comment.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listings_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/comment.ex: Permission denied
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module>
main()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main
rv = self.invoke(ctx)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed
File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 32, in <module>
shell.run(["git", "clean", "-fdx"], cwd=repository / "clients")
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 418, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1.
synthtool > Wrote metadata to clients/dlp/synth.metadata.
Synthesis failed
```
Google internal developers can see the full log [here](https://sponge/6d24954b-bcff-4c08-9c3d-47b4d58021a3).
|
1.0
|
Synthesis failed for DLP - Hello! Autosynth couldn't regenerate DLP. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth-dlp'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/dlp/synth.metadata', 'synth.py', '--']
synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py.
synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
synthtool > Failed executing git clean -fdx:
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/reviews.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/edits.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/orders.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/purchases.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/inappproducts.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/api/internalappsharingartifacts.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/connection.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundles_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_file.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/bundle.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/image.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk_binary.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listing.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/page_info.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/testers.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/season.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_details.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review_reply_result.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/tracks_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_defer_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchase.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchase.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/product_purchases_acknowledge_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_cancel_survey_result.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/timestamp.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/deobfuscation_files_upload_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_upload_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_price_change.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_deferral_info.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/subscription_purchases_acknowledge_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchase.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_delete_all_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/price.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/prorate.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk_uses_permission.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/developer_comment.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/images_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/voided_purchases_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/review.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_file.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/token_pagination.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apk.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/localized_text.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/country_targeting.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/month_day.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/reviews_reply_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/expansion_files_upload_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/externally_hosted_apk.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/apks_add_externally_hosted_request.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/track_release.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/device_metadata.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/in_app_product_listing.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/app_edit.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/internal_app_sharing_artifact.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/inappproducts_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/user_comment.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/listings_list_response.ex: Permission denied
warning: failed to remove android_publisher/lib/google_api/android_publisher/v3/model/comment.ex: Permission denied
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module>
main()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main
rv = self.invoke(ctx)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed
File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 32, in <module>
shell.run(["git", "clean", "-fdx"], cwd=repository / "clients")
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 418, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1.
synthtool > Wrote metadata to clients/dlp/synth.metadata.
Synthesis failed
```
Google internal developers can see the full log [here](https://sponge/6d24954b-bcff-4c08-9c3d-47b4d58021a3).
|
non_process
|
synthesis failed for dlp hello autosynth couldn t regenerate dlp broken heart here s the output from running synth py cloning into working repo switched to branch autosynth dlp running synthtool synthtool executing tmpfs src git autosynth working repo synth py synthtool cloning synthtool failed executing git clean fdx warning failed to remove android publisher lib google api android publisher api reviews ex permission denied warning failed to remove android publisher lib google api android publisher api edits ex permission denied warning failed to remove android publisher lib google api android publisher api orders ex permission denied warning failed to remove android publisher lib google api android publisher api purchases ex permission denied warning failed to remove android publisher lib google api android publisher api inappproducts ex permission denied warning failed to remove android publisher lib google api android publisher api internalappsharingartifacts ex permission denied warning failed to remove android publisher lib google api android publisher connection ex permission denied warning failed to remove android publisher lib google api android publisher model bundles list response ex permission denied warning failed to remove android publisher lib google api android publisher model deobfuscation file ex permission denied warning failed to remove android publisher lib google api android publisher model bundle ex permission denied warning failed to remove android publisher lib google api android publisher model image ex permission denied warning failed to remove android publisher lib google api android publisher model apk binary ex permission denied warning failed to remove android publisher lib google api android publisher model listing ex permission denied warning failed to remove android publisher lib google api android publisher model page info ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchases defer request ex permission denied warning failed to remove android publisher lib google api android publisher model testers ex permission denied warning failed to remove android publisher lib google api android publisher model track ex permission denied warning failed to remove android publisher lib google api android publisher model in app product ex permission denied warning failed to remove android publisher lib google api android publisher model season ex permission denied warning failed to remove android publisher lib google api android publisher model app details ex permission denied warning failed to remove android publisher lib google api android publisher model apks list response ex permission denied warning failed to remove android publisher lib google api android publisher model review reply result ex permission denied warning failed to remove android publisher lib google api android publisher model tracks list response ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchases defer response ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchase ex permission denied warning failed to remove android publisher lib google api android publisher model product purchase ex permission denied warning failed to remove android publisher lib google api android publisher model product purchases acknowledge request ex permission denied warning failed to remove android publisher lib google api android publisher model subscription cancel survey result ex permission denied warning failed to remove android publisher lib google api android publisher model timestamp ex permission denied warning failed to remove android publisher lib google api android publisher model deobfuscation files upload response ex permission denied warning failed to remove android publisher lib google api android publisher model images upload response ex permission denied warning failed to remove android publisher lib google api android publisher model subscription price change ex permission denied warning failed to remove android publisher lib google api android publisher model subscription deferral info ex permission denied warning failed to remove android publisher lib google api android publisher model subscription purchases acknowledge request ex permission denied warning failed to remove android publisher lib google api android publisher model reviews list response ex permission denied warning failed to remove android publisher lib google api android publisher model voided purchase ex permission denied warning failed to remove android publisher lib google api android publisher model images delete all response ex permission denied warning failed to remove android publisher lib google api android publisher model price ex permission denied warning failed to remove android publisher lib google api android publisher model reviews reply request ex permission denied warning failed to remove android publisher lib google api android publisher model prorate ex permission denied warning failed to remove android publisher lib google api android publisher model externally hosted apk uses permission ex permission denied warning failed to remove android publisher lib google api android publisher model developer comment ex permission denied warning failed to remove android publisher lib google api android publisher model images list response ex permission denied warning failed to remove android publisher lib google api android publisher model voided purchases list response ex permission denied warning failed to remove android publisher lib google api android publisher model review ex permission denied warning failed to remove android publisher lib google api android publisher model expansion file ex permission denied warning failed to remove android publisher lib google api android publisher model token pagination ex permission denied warning failed to remove android publisher lib google api android publisher model apk ex permission denied warning failed to remove android publisher lib google api android publisher model localized text ex permission denied warning failed to remove android publisher lib google api android publisher model country targeting ex permission denied warning failed to remove android publisher lib google api android publisher model month day ex permission denied warning failed to remove android publisher lib google api android publisher model apks add externally hosted response ex permission denied warning failed to remove android publisher lib google api android publisher model reviews reply response ex permission denied warning failed to remove android publisher lib google api android publisher model expansion files upload response ex permission denied warning failed to remove android publisher lib google api android publisher model externally hosted apk ex permission denied warning failed to remove android publisher lib google api android publisher model apks add externally hosted request ex permission denied warning failed to remove android publisher lib google api android publisher model track release ex permission denied warning failed to remove android publisher lib google api android publisher model device metadata ex permission denied warning failed to remove android publisher lib google api android publisher model in app product listing ex permission denied warning failed to remove android publisher lib google api android publisher model app edit ex permission denied warning failed to remove android publisher lib google api android publisher model internal app sharing artifact ex permission denied warning failed to remove android publisher lib google api android publisher model inappproducts list response ex permission denied warning failed to remove android publisher lib google api android publisher model user comment ex permission denied warning failed to remove android publisher lib google api android publisher model listings list response ex permission denied warning failed to remove android publisher lib google api android publisher model comment ex permission denied traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth env lib site packages synthtool main py line in main file tmpfs src git autosynth env lib site packages click core py line in call return self main args kwargs file tmpfs src git autosynth env lib site packages click core py line in main rv self invoke ctx file tmpfs src git autosynth env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src git autosynth env lib site packages click core py line in invoke return callback args kwargs file tmpfs src git autosynth env lib site packages synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src git autosynth working repo synth py line in shell run cwd repository clients file tmpfs src git autosynth env lib site packages synthtool shell py line in run raise exc file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthtool wrote metadata to clients dlp synth metadata synthesis failed google internal developers can see the full log
| 0
|
18,179
| 24,231,196,227
|
IssuesEvent
|
2022-09-26 18:26:49
|
google/android-fhir
|
https://api.github.com/repos/google/android-fhir
|
closed
|
Support Android Min API Level 24
|
process
|
We have not been testing API level 21 consistently in our testing matrix: https://github.com/google/android-fhir/blob/master/.github/workflows/device-tests.yml#L40
We should consider raising the min api level to 24 (released aug 2016) as we already do not guarantee support for anything lower than 24 due to our lack of tests.
|
1.0
|
Support Android Min API Level 24 - We have not been testing API level 21 consistently in our testing matrix: https://github.com/google/android-fhir/blob/master/.github/workflows/device-tests.yml#L40
We should consider raising the min api level to 24 (released aug 2016) as we already do not guarantee support for anything lower than 24 due to our lack of tests.
|
process
|
support android min api level we have not been testing api level consistently in our testing matrix we should consider raising the min api level to released aug as we already do not guarantee support for anything lower than due to our lack of tests
| 1
|
94,648
| 27,253,019,431
|
IssuesEvent
|
2023-02-22 09:32:20
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
closed
|
Godot 4 compiled with MinGW is crashing with a white screen [GCC 11.x bug fixed in 11.3]
|
bug platform:windows topic:buildsystem topic:thirdparty crash
|
### Godot version
4.0 custom build (bdcc8741e)
### System information
Windows 10 21H1
### Issue description
The editor is crashing with a white screen on startup, with or without a project. I have been tried some combinations of compilers and options:
MSVC with debug -> OK
MSVC with release_debug -> OK
MinGW with debug -> OK
MinGW with release_debug -> Crash
This issue has been introduced with the 99fe462452be44efa618e83ad9bbecd722ae6ecd commit.
<details>
<summary>Output of drmingw</summary>
```
godot.windows.opt.tools.64.exe caused an Access Violation at location 00007FF73A4C0000 DEP violation at location 00007FF73A4C0000.
AddrPC Params
00007FF73A4C0000 00000215616FA200 0000003CE61FF5A8 0000003CE61FF5B0
00007FF77C559842 0000000000400000 00007FF77CA33F03 0000000000400000 godot.windows.opt.tools.64.exe!RenderingServerDefault::RenderingServerDefault [E:/path/core/os/thread.h @ 90]
88: _FORCE_INLINE_ ID get_id() const { return id; }
89: // get the ID of the caller thread
> 90: _FORCE_INLINE_ static ID get_caller_id() { return caller_id; }
91: // get the ID of the main thread
92: _FORCE_INLINE_ static ID get_main_id() { return main_thread_id; }
00007FF77A4DE90B 0000021561579CC0 0000021560C1F690 00000215000000FF godot.windows.opt.tools.64.exe!Main::setup2 [E:/path/main/main.cpp @ 1590]
1588: /* Initialize Rendering Server */
1589:
> 1590: rendering_server = memnew(RenderingServerDefault(OS::get_singleton()->get_render_thread_mode() == OS::RENDER_SEPARATE_THREAD));
1591:
1592: rendering_server->init();
00007FF77A4E479B 0000021560C20C30 0000000000000001 000002155F450860 godot.windows.opt.tools.64.exe!Main::setup [E:/path/main/main.cpp @ 1388]
1386:
1387: if (p_second_phase) {
> 1388: return setup2();
1389: }
1390:
00007FF77A4C18A7 000002155F153296 0000003CE61FFD4C 0000000000000001 godot.windows.opt.tools.64.exe!widechar_main [E:/path/platform/windows/godot_windows.cpp @ 151]
149: TEST_MAIN_PARAM_OVERRIDE(argc, argv_utf8)
150:
> 151: Error err = Main::setup(argv_utf8[0], argc - 1, &argv_utf8[1]);
152:
153: if (err != OK) {
00007FF77A4C19A0 0000000000000000 0000000000000047 00007FF7811164D8 godot.windows.opt.tools.64.exe!_main [E:/path/platform/windows/godot_windows.cpp @ 185]
183: }
184:
> 185: result = widechar_main(argc, wc_argv);
186:
187: LocalFree(wc_argv);
00007FF77A4C13B1 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!__tmainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 321]
00007FF77A4C14E6 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!mainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 202]
00007FF8F8997034 0000000000000000 0000000000000000 0000000000000000 KERNEL32.DLL!BaseThreadInitThunk
00007FF8FA1C2651 0000000000000000 0000000000000000 0000000000000000 ntdll.dll!RtlUserThreadStart
```
</details>
<details>
<summary>Log of the crashing MinGW release_debug build</summary>
```
Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org
Using "ICU / HarfBuzz / Graphite" text server...
Vulkan API 1.2.162
ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157)
WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files.
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154)
Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080
- Vulkan multiview supported:
max view count: 32
max instances: 134217727
- Vulkan subgroup:
size: 32
stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV
supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV
quad operations in all stages
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using "winink" pen tablet driver...
```
</details>
<details>
<summary>Log of the working MinGW debug build</summary>
```
Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org
Using "ICU / HarfBuzz / Graphite" text server...
Vulkan API 1.2.162
ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157)
WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files.
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154)
Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080
- Vulkan multiview supported:
max view count: 32
max instances: 134217727
- Vulkan subgroup:
size: 32
stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV
supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV
quad operations in all stages
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using "winink" pen tablet driver...
Shader 'VoxelGiSdfShaderRD' SHA256: b6972a55781e21cfbac0dfc035381ee1f65e972c1f7e0bcc6004bfb74cc9dc9e
Shader 'ParticlesShaderRD' SHA256: cbb6a5ce1933875a8e2df4f0cde4684286dedb6737cfbcfdeba745c978a2cc45
Shader 'ParticlesCopyShaderRD' SHA256: 6bd2ea6fca4c3a51f1e9bcd5068315e63820ec68fd8596e20480174c943bb9df
Shader 'CanvasSdfShaderRD' SHA256: 692328d90ee7d404503d304e2464af3f3ac466ebd287762fd8883248f3e2a046
Shader 'SkeletonShaderRD' SHA256: 8f03a2ebc895d833120c3b2341c16e971fb1404dd3ae9e8030bc4bbf63e67fd5
Shader 'CanvasShaderRD' SHA256: a30a00a53b17f4e413dfb1fe109868063f2d67900aaeec66b0f50964108e216b
Shader 'CanvasOcclusionShaderRD' SHA256: fd08d5c92b6537852fb4b8e5fba27f28518faad41a5c21ac9bd51bf49807824c
Shader 'ClusterRenderShaderRD' SHA256: a2924afb04064a1dc42ed9d63721e7d5dd0466f245037ff85b5bef8bd251f515
Shader 'ClusterStoreShaderRD' SHA256: 64c437e6bebba9faa4766b524281310d934fd2a28f26e9a8a7834ca07f09d2dc
Shader 'ClusterDebugShaderRD' SHA256: 0f66ac91d81e6f440ffa804d473579cb5d7ac135b375dce6d120186a65b54204
Shader 'SkyShaderRD' SHA256: dd0dbf614e400cc2f58360d292f85f19237fdd9746b9a46c6fd7c50fe8a2ae43
Shader 'VoxelGiShaderRD' SHA256: 51aa13996be8fb6071160d3ea73d752edcd67ef81e1aec1ec3623a13eee94d0a
Shader 'VoxelGiDebugShaderRD' SHA256: 7fa6be49a459f8891a6e756165f2e9ca7f3ba261f53edf36c2041dc2acbf2a96
Shader 'SdfgiPreprocessShaderRD' SHA256: 056a9e6eaae9cc98093ceec6142bcb4582be0f3443e18402c1e25c293899b104
Shader 'SdfgiDirectLightShaderRD' SHA256: 0707e534eb11aeb7566a94913fe168d96b66896b440650f768f92ec6ed5cc842
Shader 'SdfgiIntegrateShaderRD' SHA256: deacba90b0c075f22f4000295f6a0b846957d2602b5d54b631c089792288e7df
Shader 'GiShaderRD' SHA256: c1d65ee198e82b396759752f26b64ae98d7bb8691aed1bc07924617cf25872c5
Shader 'SdfgiDebugShaderRD' SHA256: 5e9f1e4d567be5c04dacec2ddbddb98638086a1fe7dc1d0260f3a58e93856a0e
Shader 'SdfgiDebugProbesShaderRD' SHA256: a6a825191c0869482f513376305eb79fa16189e65068bf76a484a0e1eb602c1b
Shader 'VolumetricFogShaderRD' SHA256: bc8fa35c7d92fd0353d5458af7d19438bf0c46f8c873b1b5e45f328f988f2d34
Shader 'SceneForwardClusteredShaderRD' SHA256: 300732714cbaf090328e341566d4d056b1bf83a2786d31de56778f6c929d1987
Shader 'CopyShaderRD' SHA256: 4ea51535b5cdcba7ece20ebd074657d05cbde47d1d0e87d3c84a2a1c28aa976f
Shader 'CopyToFbShaderRD' SHA256: 6972e6ea6ece204edb2197ac92f699ad4e7c4b22cac23422dd816a2a9fc11ce3
Shader 'CubemapRoughnessShaderRD' SHA256: 083d83f629fde69ed8e4c482638f30df40144b9e05c8d8243317b1866ff97897
Shader 'TonemapShaderRD' SHA256: cd4d7156689aa910864bd876e2c439c8f2b49dd281df707f93145fec99b9bd67
Shader 'LuminanceReduceShaderRD' SHA256: f8db4abaa10d52974aef5a0443e69f8838eb05f4b4eee120c19ca3b0f8ef6353
Shader 'CubeToDpShaderRD' SHA256: 1ac5646aaebfed0f2fa62c6faa2f87981b8a5acf4748191482448d7fe516cc78
Shader 'BokehDofShaderRD' SHA256: 7ae11ecc5f3863fff19a28c9e41a9ba8016c53f54cc182daa15e6a1e080ce3c6
Shader 'SsaoDownsampleShaderRD' SHA256: 4db35346ca9c82436e993cde3466bee499a8f66f6d58e39d9bed78ee7d835e1f
Shader 'SsaoShaderRD' SHA256: db7ede5f05458aaa5e03db5e62804bcd7e3b4b3eeb556e9eb027a958a96b6bd7
Shader 'SsaoImportanceMapShaderRD' SHA256: a2305d0dc9bed6da701e58b6d6a28bb2a664aedb9de4460c390b21a7f989e8eb
Shader 'SsaoBlurShaderRD' SHA256: 9158ebb0e50201b64d1c1009317e7cf5de0792a82a4d06b393e5db5ec351b54f
Shader 'SsaoInterleaveShaderRD' SHA256: e8d5eda46653d75788db067517688755630bcb40ff6da19c5f039db86df234c3
Shader 'RoughnessLimiterShaderRD' SHA256: fe30b9757ac0d3bbb53ab48bc7e0d48cc760f8212fc723c588c30dc9466539d4
Shader 'CubemapDownsamplerShaderRD' SHA256: 8184770cb634e8d5b7f7d4e801b864caa2972a7a98181a62c196ffa7b882edc3
Shader 'CubemapFilterShaderRD' SHA256: 5c27e7fe5c0fce79c147c7a8ae6984a0a7ae2dd90af07e35a4a34934698a91ca
Shader 'SpecularMergeShaderRD' SHA256: f6dc7afac1a2715cc2f134daff1bddfd6b7ca715d655aca8cafa32798395dec2
Shader 'ScreenSpaceReflectionShaderRD' SHA256: d56b77268d5c239c3886ca85201d733c47ea7e6542969c5eb6d5350b483778ef
Shader 'ScreenSpaceReflectionFilterShaderRD' SHA256: 4cd6afe65db3d97e059daa77702faf2a6e6cccb9c552c61af6055cb5be448b5f
Shader 'ScreenSpaceReflectionScaleShaderRD' SHA256: 89c672345ad85d25a8e9cbc814aac4007cb5fcb29cb8262d0185abd23fdccb9f
Shader 'SubsurfaceScatteringShaderRD' SHA256: c633c5beb82a7102c2849bec06732e4c6c98ebfc17b013f0844089d30beaef52
Shader 'ResolveShaderRD' SHA256: d0c9281d4f856dc19c5c6d14e0b256b28ca5455371db93744c010476572dbf25
Shader 'SortShaderRD' SHA256: 2192a5f1b643af9d53a2a609c7d2a715237ea16c083317a1611af42c6845b99c
Shader 'BlitShaderRD' SHA256: 3c3477999355f834dd2a898d94df5b40ae97299bb1f5e9de2dd2b6caf46055f5
WASAPI: wFormatTag = 65534
WASAPI: nChannels = 8
WASAPI: nSamplesPerSec = 48000
WASAPI: nAvgBytesPerSec = 1536000
WASAPI: nBlockAlign = 32
WASAPI: wBitsPerSample = 32
WASAPI: cbSize = 22
WASAPI: detected 8 channels
WASAPI: audio buffer frames: 1962 calculated latency: 44ms
ERROR: Attempted to free invalid ID: 2100239007746
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
CORE API HASH: 12701809910231588744
EDITOR API HASH: 7657489752925521230
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Loaded builtin certs
EditorSettings: Save OK!
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using present mode: VK_PRESENT_MODE_FIFO_KHR
EditorSettings: Save OK!
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: 1 RID allocations of type 'N17RendererStorageRD7TextureE' were leaked at exit.
WARNING: 2 RIDs of type "Texture" were leaked.
at: finalize (drivers\vulkan\rendering_device_vulkan.cpp:9055)
StringName: 173 unclaimed string names at exit.
```
</details>
Summary of tried combinations and result of it:
| Distribution | MinGW version | GCC version | binutils version | Works |
|--------------|------------------------|-------------|------------------|-------|
| Fedora 34 | 8.0.0 | 10.3.1 | 2.34 | yes |
| Fedora 35 | 9.0.0 | 11.2.1 | 2.37 | no |
| Debian Sid | 8.0.0 | 10.2.1 | 2.37 | no |
| OpenSuse TW | 9.0.0 | 9.2.0 | 2.33 | yes |
| MSYS2 | 9.0.0.6346.6cc97775a-1 | 11.2.0 | 2.37 | no |
| Mageia 8 | 8.0.0 | 10.2.1 | 2.34 | yes<sup>[1]</sup> |
| MSYS2 | 9.0.0.6454.b4445ee52 | 11.2.0 | 2.38 | no |
| MSYS2 | 10.0.0.r0.gaa08f56da-1 | 11.2.0 | 2.38 | no |
[1] The icon from the executable was missing
### Steps to reproduce
For a not working build:
`scons platform=windows target=release_debug use_mingw=yes -j12`
For a working build:
`scons platform=windows target=debug use_mingw=yes -j12`
### Minimal reproduction project
_No response_
|
1.0
|
Godot 4 compiled with MinGW is crashing with a white screen [GCC 11.x bug fixed in 11.3] - ### Godot version
4.0 custom build (bdcc8741e)
### System information
Windows 10 21H1
### Issue description
The editor is crashing with a white screen on startup, with or without a project. I have been tried some combinations of compilers and options:
MSVC with debug -> OK
MSVC with release_debug -> OK
MinGW with debug -> OK
MinGW with release_debug -> Crash
This issue has been introduced with the 99fe462452be44efa618e83ad9bbecd722ae6ecd commit.
<details>
<summary>Output of drmingw</summary>
```
godot.windows.opt.tools.64.exe caused an Access Violation at location 00007FF73A4C0000 DEP violation at location 00007FF73A4C0000.
AddrPC Params
00007FF73A4C0000 00000215616FA200 0000003CE61FF5A8 0000003CE61FF5B0
00007FF77C559842 0000000000400000 00007FF77CA33F03 0000000000400000 godot.windows.opt.tools.64.exe!RenderingServerDefault::RenderingServerDefault [E:/path/core/os/thread.h @ 90]
88: _FORCE_INLINE_ ID get_id() const { return id; }
89: // get the ID of the caller thread
> 90: _FORCE_INLINE_ static ID get_caller_id() { return caller_id; }
91: // get the ID of the main thread
92: _FORCE_INLINE_ static ID get_main_id() { return main_thread_id; }
00007FF77A4DE90B 0000021561579CC0 0000021560C1F690 00000215000000FF godot.windows.opt.tools.64.exe!Main::setup2 [E:/path/main/main.cpp @ 1590]
1588: /* Initialize Rendering Server */
1589:
> 1590: rendering_server = memnew(RenderingServerDefault(OS::get_singleton()->get_render_thread_mode() == OS::RENDER_SEPARATE_THREAD));
1591:
1592: rendering_server->init();
00007FF77A4E479B 0000021560C20C30 0000000000000001 000002155F450860 godot.windows.opt.tools.64.exe!Main::setup [E:/path/main/main.cpp @ 1388]
1386:
1387: if (p_second_phase) {
> 1388: return setup2();
1389: }
1390:
00007FF77A4C18A7 000002155F153296 0000003CE61FFD4C 0000000000000001 godot.windows.opt.tools.64.exe!widechar_main [E:/path/platform/windows/godot_windows.cpp @ 151]
149: TEST_MAIN_PARAM_OVERRIDE(argc, argv_utf8)
150:
> 151: Error err = Main::setup(argv_utf8[0], argc - 1, &argv_utf8[1]);
152:
153: if (err != OK) {
00007FF77A4C19A0 0000000000000000 0000000000000047 00007FF7811164D8 godot.windows.opt.tools.64.exe!_main [E:/path/platform/windows/godot_windows.cpp @ 185]
183: }
184:
> 185: result = widechar_main(argc, wc_argv);
186:
187: LocalFree(wc_argv);
00007FF77A4C13B1 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!__tmainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 321]
00007FF77A4C14E6 0000000000000000 0000000000000000 0000000000000000 godot.windows.opt.tools.64.exe!mainCRTStartup [C:/_/M/mingw-w64-crt-git/src/mingw-w64/mingw-w64-crt/crt/crtexe.c @ 202]
00007FF8F8997034 0000000000000000 0000000000000000 0000000000000000 KERNEL32.DLL!BaseThreadInitThunk
00007FF8FA1C2651 0000000000000000 0000000000000000 0000000000000000 ntdll.dll!RtlUserThreadStart
```
</details>
<details>
<summary>Log of the crashing MinGW release_debug build</summary>
```
Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org
Using "ICU / HarfBuzz / Graphite" text server...
Vulkan API 1.2.162
ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157)
WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files.
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 2934791902880
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154)
Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080
- Vulkan multiview supported:
max view count: 32
max instances: 134217727
- Vulkan subgroup:
size: 32
stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV
supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV
quad operations in all stages
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using "winink" pen tablet driver...
```
</details>
<details>
<summary>Log of the working MinGW debug build</summary>
```
Godot Engine v4.0.dev.custom_build.bdcc8741e - https://godotengine.org
Using "ICU / HarfBuzz / Graphite" text server...
Vulkan API 1.2.162
ERROR: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
loader_get_json: Failed to open JSON file C:\ProgramData\GOG.com\Galaxy\redists\overlay\injected\galaxy_overlay_vklayer_x64.json
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:157)
WARNING: GENERAL - Message Id Number: 0 | Message Id Name: Loader Message
ReadDataFilesInRegistry: Registry lookup failed to get layer manifest files.
Objects - 1
Object[0] - VK_OBJECT_TYPE_INSTANCE, Handle 1835716830080
at: _debug_messenger_callback (drivers\vulkan\vulkan_context.cpp:154)
Using Vulkan Device #0: NVIDIA - NVIDIA GeForce RTX 2080
- Vulkan multiview supported:
max view count: 32
max instances: 134217727
- Vulkan subgroup:
size: 32
stages: STAGE_VERTEX, STAGE_TESSELLATION_CONTROL, STAGE_TESSELLATION_EVALUATION, STAGE_GEOMETRY, STAGE_FRAGMENT, STAGE_COMPUTE, STAGE_RAYGEN_KHR, STAGE_ANY_HIT_KHR, STAGE_CLOSEST_HIT_KHR, STAGE_MISS_KHR, STAGE_INTERSECTION_KHR, STAGE_CALLABLE_KHR, STAGE_TASK_NV, STAGE_MESH_NV
supported ops: FEATURE_BASIC, FEATURE_VOTE, FEATURE_ARITHMETIC, FEATURE_BALLOT, FEATURE_SHUFFLE, FEATURE_SHUFFLE_RELATIVE, FEATURE_CLUSTERED, FEATURE_QUAD, FEATURE_PARTITIONED_NV
quad operations in all stages
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using "winink" pen tablet driver...
Shader 'VoxelGiSdfShaderRD' SHA256: b6972a55781e21cfbac0dfc035381ee1f65e972c1f7e0bcc6004bfb74cc9dc9e
Shader 'ParticlesShaderRD' SHA256: cbb6a5ce1933875a8e2df4f0cde4684286dedb6737cfbcfdeba745c978a2cc45
Shader 'ParticlesCopyShaderRD' SHA256: 6bd2ea6fca4c3a51f1e9bcd5068315e63820ec68fd8596e20480174c943bb9df
Shader 'CanvasSdfShaderRD' SHA256: 692328d90ee7d404503d304e2464af3f3ac466ebd287762fd8883248f3e2a046
Shader 'SkeletonShaderRD' SHA256: 8f03a2ebc895d833120c3b2341c16e971fb1404dd3ae9e8030bc4bbf63e67fd5
Shader 'CanvasShaderRD' SHA256: a30a00a53b17f4e413dfb1fe109868063f2d67900aaeec66b0f50964108e216b
Shader 'CanvasOcclusionShaderRD' SHA256: fd08d5c92b6537852fb4b8e5fba27f28518faad41a5c21ac9bd51bf49807824c
Shader 'ClusterRenderShaderRD' SHA256: a2924afb04064a1dc42ed9d63721e7d5dd0466f245037ff85b5bef8bd251f515
Shader 'ClusterStoreShaderRD' SHA256: 64c437e6bebba9faa4766b524281310d934fd2a28f26e9a8a7834ca07f09d2dc
Shader 'ClusterDebugShaderRD' SHA256: 0f66ac91d81e6f440ffa804d473579cb5d7ac135b375dce6d120186a65b54204
Shader 'SkyShaderRD' SHA256: dd0dbf614e400cc2f58360d292f85f19237fdd9746b9a46c6fd7c50fe8a2ae43
Shader 'VoxelGiShaderRD' SHA256: 51aa13996be8fb6071160d3ea73d752edcd67ef81e1aec1ec3623a13eee94d0a
Shader 'VoxelGiDebugShaderRD' SHA256: 7fa6be49a459f8891a6e756165f2e9ca7f3ba261f53edf36c2041dc2acbf2a96
Shader 'SdfgiPreprocessShaderRD' SHA256: 056a9e6eaae9cc98093ceec6142bcb4582be0f3443e18402c1e25c293899b104
Shader 'SdfgiDirectLightShaderRD' SHA256: 0707e534eb11aeb7566a94913fe168d96b66896b440650f768f92ec6ed5cc842
Shader 'SdfgiIntegrateShaderRD' SHA256: deacba90b0c075f22f4000295f6a0b846957d2602b5d54b631c089792288e7df
Shader 'GiShaderRD' SHA256: c1d65ee198e82b396759752f26b64ae98d7bb8691aed1bc07924617cf25872c5
Shader 'SdfgiDebugShaderRD' SHA256: 5e9f1e4d567be5c04dacec2ddbddb98638086a1fe7dc1d0260f3a58e93856a0e
Shader 'SdfgiDebugProbesShaderRD' SHA256: a6a825191c0869482f513376305eb79fa16189e65068bf76a484a0e1eb602c1b
Shader 'VolumetricFogShaderRD' SHA256: bc8fa35c7d92fd0353d5458af7d19438bf0c46f8c873b1b5e45f328f988f2d34
Shader 'SceneForwardClusteredShaderRD' SHA256: 300732714cbaf090328e341566d4d056b1bf83a2786d31de56778f6c929d1987
Shader 'CopyShaderRD' SHA256: 4ea51535b5cdcba7ece20ebd074657d05cbde47d1d0e87d3c84a2a1c28aa976f
Shader 'CopyToFbShaderRD' SHA256: 6972e6ea6ece204edb2197ac92f699ad4e7c4b22cac23422dd816a2a9fc11ce3
Shader 'CubemapRoughnessShaderRD' SHA256: 083d83f629fde69ed8e4c482638f30df40144b9e05c8d8243317b1866ff97897
Shader 'TonemapShaderRD' SHA256: cd4d7156689aa910864bd876e2c439c8f2b49dd281df707f93145fec99b9bd67
Shader 'LuminanceReduceShaderRD' SHA256: f8db4abaa10d52974aef5a0443e69f8838eb05f4b4eee120c19ca3b0f8ef6353
Shader 'CubeToDpShaderRD' SHA256: 1ac5646aaebfed0f2fa62c6faa2f87981b8a5acf4748191482448d7fe516cc78
Shader 'BokehDofShaderRD' SHA256: 7ae11ecc5f3863fff19a28c9e41a9ba8016c53f54cc182daa15e6a1e080ce3c6
Shader 'SsaoDownsampleShaderRD' SHA256: 4db35346ca9c82436e993cde3466bee499a8f66f6d58e39d9bed78ee7d835e1f
Shader 'SsaoShaderRD' SHA256: db7ede5f05458aaa5e03db5e62804bcd7e3b4b3eeb556e9eb027a958a96b6bd7
Shader 'SsaoImportanceMapShaderRD' SHA256: a2305d0dc9bed6da701e58b6d6a28bb2a664aedb9de4460c390b21a7f989e8eb
Shader 'SsaoBlurShaderRD' SHA256: 9158ebb0e50201b64d1c1009317e7cf5de0792a82a4d06b393e5db5ec351b54f
Shader 'SsaoInterleaveShaderRD' SHA256: e8d5eda46653d75788db067517688755630bcb40ff6da19c5f039db86df234c3
Shader 'RoughnessLimiterShaderRD' SHA256: fe30b9757ac0d3bbb53ab48bc7e0d48cc760f8212fc723c588c30dc9466539d4
Shader 'CubemapDownsamplerShaderRD' SHA256: 8184770cb634e8d5b7f7d4e801b864caa2972a7a98181a62c196ffa7b882edc3
Shader 'CubemapFilterShaderRD' SHA256: 5c27e7fe5c0fce79c147c7a8ae6984a0a7ae2dd90af07e35a4a34934698a91ca
Shader 'SpecularMergeShaderRD' SHA256: f6dc7afac1a2715cc2f134daff1bddfd6b7ca715d655aca8cafa32798395dec2
Shader 'ScreenSpaceReflectionShaderRD' SHA256: d56b77268d5c239c3886ca85201d733c47ea7e6542969c5eb6d5350b483778ef
Shader 'ScreenSpaceReflectionFilterShaderRD' SHA256: 4cd6afe65db3d97e059daa77702faf2a6e6cccb9c552c61af6055cb5be448b5f
Shader 'ScreenSpaceReflectionScaleShaderRD' SHA256: 89c672345ad85d25a8e9cbc814aac4007cb5fcb29cb8262d0185abd23fdccb9f
Shader 'SubsurfaceScatteringShaderRD' SHA256: c633c5beb82a7102c2849bec06732e4c6c98ebfc17b013f0844089d30beaef52
Shader 'ResolveShaderRD' SHA256: d0c9281d4f856dc19c5c6d14e0b256b28ca5455371db93744c010476572dbf25
Shader 'SortShaderRD' SHA256: 2192a5f1b643af9d53a2a609c7d2a715237ea16c083317a1611af42c6845b99c
Shader 'BlitShaderRD' SHA256: 3c3477999355f834dd2a898d94df5b40ae97299bb1f5e9de2dd2b6caf46055f5
WASAPI: wFormatTag = 65534
WASAPI: nChannels = 8
WASAPI: nSamplesPerSec = 48000
WASAPI: nAvgBytesPerSec = 1536000
WASAPI: nBlockAlign = 32
WASAPI: wBitsPerSample = 32
WASAPI: cbSize = 22
WASAPI: detected 8 channels
WASAPI: audio buffer frames: 1962 calculated latency: 44ms
ERROR: Attempted to free invalid ID: 2100239007746
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
CORE API HASH: 12701809910231588744
EDITOR API HASH: 7657489752925521230
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Loaded builtin certs
EditorSettings: Save OK!
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using present mode: VK_PRESENT_MODE_FIFO_KHR
Using present mode: VK_PRESENT_MODE_FIFO_KHR
EditorSettings: Save OK!
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: Attempted to free invalid ID: 0
at: _free_internal (drivers\vulkan\rendering_device_vulkan.cpp:8313)
ERROR: 1 RID allocations of type 'N17RendererStorageRD7TextureE' were leaked at exit.
WARNING: 2 RIDs of type "Texture" were leaked.
at: finalize (drivers\vulkan\rendering_device_vulkan.cpp:9055)
StringName: 173 unclaimed string names at exit.
```
</details>
Summary of tried combinations and result of it:
| Distribution | MinGW version | GCC version | binutils version | Works |
|--------------|------------------------|-------------|------------------|-------|
| Fedora 34 | 8.0.0 | 10.3.1 | 2.34 | yes |
| Fedora 35 | 9.0.0 | 11.2.1 | 2.37 | no |
| Debian Sid | 8.0.0 | 10.2.1 | 2.37 | no |
| OpenSuse TW | 9.0.0 | 9.2.0 | 2.33 | yes |
| MSYS2 | 9.0.0.6346.6cc97775a-1 | 11.2.0 | 2.37 | no |
| Mageia 8 | 8.0.0 | 10.2.1 | 2.34 | yes<sup>[1]</sup> |
| MSYS2 | 9.0.0.6454.b4445ee52 | 11.2.0 | 2.38 | no |
| MSYS2 | 10.0.0.r0.gaa08f56da-1 | 11.2.0 | 2.38 | no |
[1] The icon from the executable was missing
### Steps to reproduce
For a not working build:
`scons platform=windows target=release_debug use_mingw=yes -j12`
For a working build:
`scons platform=windows target=debug use_mingw=yes -j12`
### Minimal reproduction project
_No response_
|
non_process
|
godot compiled with mingw is crashing with a white screen godot version custom build system information windows issue description the editor is crashing with a white screen on startup with or without a project i have been tried some combinations of compilers and options msvc with debug ok msvc with release debug ok mingw with debug ok mingw with release debug crash this issue has been introduced with the commit output of drmingw godot windows opt tools exe caused an access violation at location dep violation at location addrpc params godot windows opt tools exe renderingserverdefault renderingserverdefault force inline id get id const return id get the id of the caller thread force inline static id get caller id return caller id get the id of the main thread force inline static id get main id return main thread id godot windows opt tools exe main initialize rendering server rendering server memnew renderingserverdefault os get singleton get render thread mode os render separate thread rendering server init godot windows opt tools exe main setup if p second phase return godot windows opt tools exe widechar main test main param override argc argv error err main setup argv argc argv if err ok godot windows opt tools exe main result widechar main argc wc argv localfree wc argv godot windows opt tools exe tmaincrtstartup godot windows opt tools exe maincrtstartup dll basethreadinitthunk ntdll dll rtluserthreadstart log of the crashing mingw release debug build godot engine dev custom build using icu harfbuzz graphite text server vulkan api error general message id number message id name loader message loader get json failed to open json file c programdata gog com galaxy redists overlay injected galaxy overlay vklayer json objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp warning general message id number message id name loader message readdatafilesinregistry registry lookup failed to get layer manifest files objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp using vulkan device nvidia nvidia geforce rtx vulkan multiview supported max view count max instances vulkan subgroup size stages stage vertex stage tessellation control stage tessellation evaluation stage geometry stage fragment stage compute stage raygen khr stage any hit khr stage closest hit khr stage miss khr stage intersection khr stage callable khr stage task nv stage mesh nv supported ops feature basic feature vote feature arithmetic feature ballot feature shuffle feature shuffle relative feature clustered feature quad feature partitioned nv quad operations in all stages using present mode vk present mode fifo khr using winink pen tablet driver log of the working mingw debug build godot engine dev custom build using icu harfbuzz graphite text server vulkan api error general message id number message id name loader message loader get json failed to open json file c programdata gog com galaxy redists overlay injected galaxy overlay vklayer json objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp warning general message id number message id name loader message readdatafilesinregistry registry lookup failed to get layer manifest files objects object vk object type instance handle at debug messenger callback drivers vulkan vulkan context cpp using vulkan device nvidia nvidia geforce rtx vulkan multiview supported max view count max instances vulkan subgroup size stages stage vertex stage tessellation control stage tessellation evaluation stage geometry stage fragment stage compute stage raygen khr stage any hit khr stage closest hit khr stage miss khr stage intersection khr stage callable khr stage task nv stage mesh nv supported ops feature basic feature vote feature arithmetic feature ballot feature shuffle feature shuffle relative feature clustered feature quad feature partitioned nv quad operations in all stages using present mode vk present mode fifo khr using winink pen tablet driver shader voxelgisdfshaderrd shader particlesshaderrd shader particlescopyshaderrd shader canvassdfshaderrd shader skeletonshaderrd shader canvasshaderrd shader canvasocclusionshaderrd shader clusterrendershaderrd shader clusterstoreshaderrd shader clusterdebugshaderrd shader skyshaderrd shader voxelgishaderrd shader voxelgidebugshaderrd shader sdfgipreprocessshaderrd shader sdfgidirectlightshaderrd shader sdfgiintegrateshaderrd shader gishaderrd shader sdfgidebugshaderrd shader sdfgidebugprobesshaderrd shader volumetricfogshaderrd shader sceneforwardclusteredshaderrd shader copyshaderrd shader copytofbshaderrd shader cubemaproughnessshaderrd shader tonemapshaderrd shader luminancereduceshaderrd shader cubetodpshaderrd shader bokehdofshaderrd shader ssaodownsampleshaderrd shader ssaoshaderrd shader ssaoimportancemapshaderrd shader ssaoblurshaderrd shader ssaointerleaveshaderrd shader roughnesslimitershaderrd shader cubemapdownsamplershaderrd shader cubemapfiltershaderrd shader specularmergeshaderrd shader screenspacereflectionshaderrd shader screenspacereflectionfiltershaderrd shader screenspacereflectionscaleshaderrd shader subsurfacescatteringshaderrd shader resolveshaderrd shader sortshaderrd shader blitshaderrd wasapi wformattag wasapi nchannels wasapi nsamplespersec wasapi navgbytespersec wasapi nblockalign wasapi wbitspersample wasapi cbsize wasapi detected channels wasapi audio buffer frames calculated latency error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp core api hash editor api hash using present mode vk present mode fifo khr loaded builtin certs editorsettings save ok using present mode vk present mode fifo khr using present mode vk present mode fifo khr using present mode vk present mode fifo khr editorsettings save ok error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error attempted to free invalid id at free internal drivers vulkan rendering device vulkan cpp error rid allocations of type were leaked at exit warning rids of type texture were leaked at finalize drivers vulkan rendering device vulkan cpp stringname unclaimed string names at exit summary of tried combinations and result of it distribution mingw version gcc version binutils version works fedora yes fedora no debian sid no opensuse tw yes no mageia yes no no the icon from the executable was missing steps to reproduce for a not working build scons platform windows target release debug use mingw yes for a working build scons platform windows target debug use mingw yes minimal reproduction project no response
| 0
|
17,578
| 23,389,698,328
|
IssuesEvent
|
2022-08-11 16:35:22
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
closed
|
New `javascript` transform
|
needs: approval domain: transforms type: feature domain: processing
|
Recently Fabrice Bellard, creator of QEMU and FFMPEG, released [QuickJS](https://bellard.org/quickjs/) embeddable JavaScript engine.
It is MIT licensed and its compiled size is just a few hundred kilobytes.
So it might be a good idea to add `javascript` transform, similar to `lua` transform.
|
1.0
|
New `javascript` transform - Recently Fabrice Bellard, creator of QEMU and FFMPEG, released [QuickJS](https://bellard.org/quickjs/) embeddable JavaScript engine.
It is MIT licensed and its compiled size is just a few hundred kilobytes.
So it might be a good idea to add `javascript` transform, similar to `lua` transform.
|
process
|
new javascript transform recently fabrice bellard creator of qemu and ffmpeg released embeddable javascript engine it is mit licensed and its compiled size is just a few hundred kilobytes so it might be a good idea to add javascript transform similar to lua transform
| 1
|
18,132
| 24,171,033,799
|
IssuesEvent
|
2022-09-22 19:14:25
|
ORNL-AMO/AMO-Tools-Suite
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite
|
closed
|
Gas Flue Gas Edit
|
bug Process Heating
|
Issue overview
--------------
GasFlueGasMaterial.cpp
line 110
Change to...
const double combAirMoisture = combAirMoisturePerc <= 0.009 ? 0 : ((combAirMoisturePerc - 0.009) * (stoichAir * (1 + exsAir)) * 0.0763);
|
1.0
|
Gas Flue Gas Edit - Issue overview
--------------
GasFlueGasMaterial.cpp
line 110
Change to...
const double combAirMoisture = combAirMoisturePerc <= 0.009 ? 0 : ((combAirMoisturePerc - 0.009) * (stoichAir * (1 + exsAir)) * 0.0763);
|
process
|
gas flue gas edit issue overview gasfluegasmaterial cpp line change to const double combairmoisture combairmoistureperc combairmoistureperc stoichair exsair
| 1
|
19,404
| 25,544,499,540
|
IssuesEvent
|
2022-11-29 17:38:00
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process: spawn incorrect error
|
child_process
|
### Version
v18.5.0
### Platform
5.4.0-88-generic #99-Ubuntu SMP Thu Sep 23 17:29:00 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
### Subsystem
_No response_
### What steps will reproduce the bug?
```js
const spawn = require('child_process').spawn;
const cwd = magicFunction() // function return failed path: '/tmp1' - when debugging is not immediately visible
const cp = spawn('ls', [], { cwd });
cp.stdout.on('data', function(data) {
console.log(data.toString());
});
cp.stderr.on('data', function(data) {
console.log(data.toString());
});
cp.on('exit', function(code) {});
```
### How often does it reproduce? Is there a required condition?
_No response_
### What is the expected behavior?
Error: spawn cwd ENOENT
### What do you see instead?
Error: spawn ls ENOENT
### Additional information
Hi, when calling system commands error is misleading, literally, spawn can't find ls,
actually the problem is in cwd.
It seemed that the problem is with the first argument `ls` --> `/bin/bash` and two argument `[]` --> `['-c', 'ls']` or systems environment OS PATH but not cwd
|
1.0
|
child_process: spawn incorrect error - ### Version
v18.5.0
### Platform
5.4.0-88-generic #99-Ubuntu SMP Thu Sep 23 17:29:00 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
### Subsystem
_No response_
### What steps will reproduce the bug?
```js
const spawn = require('child_process').spawn;
const cwd = magicFunction() // function return failed path: '/tmp1' - when debugging is not immediately visible
const cp = spawn('ls', [], { cwd });
cp.stdout.on('data', function(data) {
console.log(data.toString());
});
cp.stderr.on('data', function(data) {
console.log(data.toString());
});
cp.on('exit', function(code) {});
```
### How often does it reproduce? Is there a required condition?
_No response_
### What is the expected behavior?
Error: spawn cwd ENOENT
### What do you see instead?
Error: spawn ls ENOENT
### Additional information
Hi, when calling system commands error is misleading, literally, spawn can't find ls,
actually the problem is in cwd.
It seemed that the problem is with the first argument `ls` --> `/bin/bash` and two argument `[]` --> `['-c', 'ls']` or systems environment OS PATH but not cwd
|
process
|
child process spawn incorrect error version platform generic ubuntu smp thu sep utc gnu linux subsystem no response what steps will reproduce the bug js const spawn require child process spawn const cwd magicfunction function return failed path when debugging is not immediately visible const cp spawn ls cwd cp stdout on data function data console log data tostring cp stderr on data function data console log data tostring cp on exit function code how often does it reproduce is there a required condition no response what is the expected behavior error spawn cwd enoent what do you see instead error spawn ls enoent additional information hi when calling system commands error is misleading literally spawn can t find ls actually the problem is in cwd it seemed that the problem is with the first argument ls bin bash and two argument or systems environment os path but not cwd
| 1
|
14,008
| 16,814,616,075
|
IssuesEvent
|
2021-06-17 05:24:05
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
GitLab source control integration?
|
Pri2 automation/svc awaiting-product-team-response cxp process-automation/subsvc product-question triaged
|
[
Hi, Is there a way to use GitLab for SCI with Azure Automation?
]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
GitLab source control integration? -
[
Hi, Is there a way to use GitLab for SCI with Azure Automation?
]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
gitlab source control integration hi is there a way to use gitlab for sci with azure automation document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
5,328
| 8,142,119,958
|
IssuesEvent
|
2018-08-21 06:16:49
|
Madek/madek
|
https://api.github.com/repos/Madek/madek
|
opened
|
Tests Batch Editing
|
Batch process
|
- I would expect the metatags to be sorted alphabetically

- If I press "anwenden", then I do not see any kind of notice on the tile itself. I would like to see a change in colour or the button "anwenden" should vanish.

- Can you make the popup bigger, so the Derivates is not on a single line?

- It would be nice, if the amount of updated media entries would be added to the green message (Erfolgsmeldung) and if the message would not vanish

- If the filter is opened, it is hard to differentiate which is the filter and which are the fields to pick.

- Please write "Medien Einträgen" together "Medieneinträgen"

|
1.0
|
Tests Batch Editing - - I would expect the metatags to be sorted alphabetically

- If I press "anwenden", then I do not see any kind of notice on the tile itself. I would like to see a change in colour or the button "anwenden" should vanish.

- Can you make the popup bigger, so the Derivates is not on a single line?

- It would be nice, if the amount of updated media entries would be added to the green message (Erfolgsmeldung) and if the message would not vanish

- If the filter is opened, it is hard to differentiate which is the filter and which are the fields to pick.

- Please write "Medien Einträgen" together "Medieneinträgen"

|
process
|
tests batch editing i would expect the metatags to be sorted alphabetically if i press anwenden then i do not see any kind of notice on the tile itself i would like to see a change in colour or the button anwenden should vanish can you make the popup bigger so the derivates is not on a single line it would be nice if the amount of updated media entries would be added to the green message erfolgsmeldung and if the message would not vanish if the filter is opened it is hard to differentiate which is the filter and which are the fields to pick please write medien einträgen together medieneinträgen
| 1
|
495,202
| 14,273,523,211
|
IssuesEvent
|
2020-11-21 22:07:30
|
tqZK/harc_game_web
|
https://api.github.com/repos/tqZK/harc_game_web
|
closed
|
Add `category` field to Task model
|
backend base-version good first issue priority
|
Also, in view with list of tasks, divide them into categories
|
1.0
|
Add `category` field to Task model - Also, in view with list of tasks, divide them into categories
|
non_process
|
add category field to task model also in view with list of tasks divide them into categories
| 0
|
9,751
| 12,737,044,163
|
IssuesEvent
|
2020-06-25 18:01:12
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
PreProcessor files for easy Progmatic PreProcessor Definitions
|
area-System.Diagnostics.Process untriaged
|
so, this is a feature request, that might as well be for MSBuild as a whole. but no clue where to post that so....
The Problem: You have a library that you want to make possible to disable or remove certain features that you wouldn't need in a safe and practical way And you do this via PreProcessor Directives but if you wish to take it up a notch and make large libraries with multiple disablable portions you quickly run into the issue of preprocessors taking up a large portion of every file and being mostly unreadable or you needing to define multiple PreProcessors combine this with open source and having the end-user define what they need and what they don't you have an almost impossible problem. Currently only remedied via custom or cumbersome PreCompilers or Pre-Compile Processors which just add to the complexity of a project.
The Suggestion: a new file type for MSBuild *.ppf (Pre Processor File) a file that just consists of preprocessor directives, that will get processed and the resulting state will then constitute the global preprocessing state, IE if you have something like this.
[MyPreprocessor.pff]
```
#if Custom_Disable_Directive1
#define Disabled_Internal_Something_Directive
#if Conflicting_Directive
#error X depends on Y, You must choose
#endif
#endif
```
the code above would be run before processing all the other files and all defined processors added to the global store to act across all cs files.
|
1.0
|
PreProcessor files for easy Progmatic PreProcessor Definitions - so, this is a feature request, that might as well be for MSBuild as a whole. but no clue where to post that so....
The Problem: You have a library that you want to make possible to disable or remove certain features that you wouldn't need in a safe and practical way And you do this via PreProcessor Directives but if you wish to take it up a notch and make large libraries with multiple disablable portions you quickly run into the issue of preprocessors taking up a large portion of every file and being mostly unreadable or you needing to define multiple PreProcessors combine this with open source and having the end-user define what they need and what they don't you have an almost impossible problem. Currently only remedied via custom or cumbersome PreCompilers or Pre-Compile Processors which just add to the complexity of a project.
The Suggestion: a new file type for MSBuild *.ppf (Pre Processor File) a file that just consists of preprocessor directives, that will get processed and the resulting state will then constitute the global preprocessing state, IE if you have something like this.
[MyPreprocessor.pff]
```
#if Custom_Disable_Directive1
#define Disabled_Internal_Something_Directive
#if Conflicting_Directive
#error X depends on Y, You must choose
#endif
#endif
```
the code above would be run before processing all the other files and all defined processors added to the global store to act across all cs files.
|
process
|
preprocessor files for easy progmatic preprocessor definitions so this is a feature request that might as well be for msbuild as a whole but no clue where to post that so the problem you have a library that you want to make possible to disable or remove certain features that you wouldn t need in a safe and practical way and you do this via preprocessor directives but if you wish to take it up a notch and make large libraries with multiple disablable portions you quickly run into the issue of preprocessors taking up a large portion of every file and being mostly unreadable or you needing to define multiple preprocessors combine this with open source and having the end user define what they need and what they don t you have an almost impossible problem currently only remedied via custom or cumbersome precompilers or pre compile processors which just add to the complexity of a project the suggestion a new file type for msbuild ppf pre processor file a file that just consists of preprocessor directives that will get processed and the resulting state will then constitute the global preprocessing state ie if you have something like this if custom disable define disabled internal something directive if conflicting directive error x depends on y you must choose endif endif the code above would be run before processing all the other files and all defined processors added to the global store to act across all cs files
| 1
|
131,746
| 10,708,674,656
|
IssuesEvent
|
2019-10-24 20:14:15
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Fix iOS text field tests
|
a: tests ☸ platform-web
|
There are few tests in `text_field_test.dart` that cover iOS-specific behavior. Those are currently failing.
|
1.0
|
Fix iOS text field tests - There are few tests in `text_field_test.dart` that cover iOS-specific behavior. Those are currently failing.
|
non_process
|
fix ios text field tests there are few tests in text field test dart that cover ios specific behavior those are currently failing
| 0
|
336,282
| 10,179,729,915
|
IssuesEvent
|
2019-08-09 08:32:25
|
mozilla/addons-server
|
https://api.github.com/repos/mozilla/addons-server
|
closed
|
Re-sort content review queue by most recently updated
|
component: reviewer tools priority: p3
|
Currently the content review queue is sorted by "oldest to most recent" updates.
We'd like to flip it to be sorted from "most recent to oldest" updates for the August 8 push.
|
1.0
|
Re-sort content review queue by most recently updated - Currently the content review queue is sorted by "oldest to most recent" updates.
We'd like to flip it to be sorted from "most recent to oldest" updates for the August 8 push.
|
non_process
|
re sort content review queue by most recently updated currently the content review queue is sorted by oldest to most recent updates we d like to flip it to be sorted from most recent to oldest updates for the august push
| 0
|
14,047
| 16,851,695,381
|
IssuesEvent
|
2021-06-20 16:38:45
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
segfault in denoise (profiled) / nlmeans
|
bug: pending reproduce: random scope: image processing
|
**Describe the bug/issue**
segfault (backtrace at https://pastebin.com/skhGKeh4) while editing raw from https://discuss.pixls.us/t/hockey-game-under-lights-20k-iso-on-m4-3/25502.
Analysis on IRC:
```
[10:10] <@johnny_bit> darktable/src/common/nlmeans_core.c:694 caused the app to go into sigsegv
...
[10:12] <@johnny_bit> it looks like the index i might go out of range for col_sums array
...
[10:42] <@johnny_bit> You can raise an issue with that stacktrace and ping ralfbrown on it because it's part of the code he's now most "accustomed" with, adding the info we've managed to "guess" here :)
```
**To Reproduce**
Edit settings of _denoise (profiled)_ in NL-means mode (does not happen immediately, probably hard to reproduce)
**Expected behavior**
No crash
**Which commit introduced the error**
I don't know, I hardly ever use NL-means.
**Platform**
* darktable version : 864e22daf
* OS : Linux - kernel 5.11.0-18
* Linux - Distro : KUbuntu 21.04
* Memory : 64 GM
* Graphics card : Nvidia 1060/6GB
* Graphics driver : 460.80-0ubuntu0.21.04.2
* OpenCL installed : yes
* OpenCL activated : yes
* Xorg : xserver-xorg 1:7.7+22ubuntu1; xserver-xorg-core 2:1.20.11-1ubuntu1
* Desktop : KDE
* GTK+ : 3.24.25-1ubuntu4.1
* gcc : 10.3.0
* cflags :
* CMAKE_BUILD_TYPE : RelWithDebInfo
|
1.0
|
segfault in denoise (profiled) / nlmeans - **Describe the bug/issue**
segfault (backtrace at https://pastebin.com/skhGKeh4) while editing raw from https://discuss.pixls.us/t/hockey-game-under-lights-20k-iso-on-m4-3/25502.
Analysis on IRC:
```
[10:10] <@johnny_bit> darktable/src/common/nlmeans_core.c:694 caused the app to go into sigsegv
...
[10:12] <@johnny_bit> it looks like the index i might go out of range for col_sums array
...
[10:42] <@johnny_bit> You can raise an issue with that stacktrace and ping ralfbrown on it because it's part of the code he's now most "accustomed" with, adding the info we've managed to "guess" here :)
```
**To Reproduce**
Edit settings of _denoise (profiled)_ in NL-means mode (does not happen immediately, probably hard to reproduce)
**Expected behavior**
No crash
**Which commit introduced the error**
I don't know, I hardly ever use NL-means.
**Platform**
* darktable version : 864e22daf
* OS : Linux - kernel 5.11.0-18
* Linux - Distro : KUbuntu 21.04
* Memory : 64 GM
* Graphics card : Nvidia 1060/6GB
* Graphics driver : 460.80-0ubuntu0.21.04.2
* OpenCL installed : yes
* OpenCL activated : yes
* Xorg : xserver-xorg 1:7.7+22ubuntu1; xserver-xorg-core 2:1.20.11-1ubuntu1
* Desktop : KDE
* GTK+ : 3.24.25-1ubuntu4.1
* gcc : 10.3.0
* cflags :
* CMAKE_BUILD_TYPE : RelWithDebInfo
|
process
|
segfault in denoise profiled nlmeans describe the bug issue segfault backtrace at while editing raw from analysis on irc darktable src common nlmeans core c caused the app to go into sigsegv it looks like the index i might go out of range for col sums array you can raise an issue with that stacktrace and ping ralfbrown on it because it s part of the code he s now most accustomed with adding the info we ve managed to guess here to reproduce edit settings of denoise profiled in nl means mode does not happen immediately probably hard to reproduce expected behavior no crash which commit introduced the error i don t know i hardly ever use nl means platform darktable version os linux kernel linux distro kubuntu memory gm graphics card nvidia graphics driver opencl installed yes opencl activated yes xorg xserver xorg xserver xorg core desktop kde gtk gcc cflags cmake build type relwithdebinfo
| 1
|
10,724
| 8,696,995,781
|
IssuesEvent
|
2018-12-04 19:02:08
|
webhintio/hint
|
https://api.github.com/repos/webhintio/hint
|
closed
|
Provide more contribution related details
|
area:infrastructure
|
- [x] Add `ISSUE_TEMPLATE.md`.
- [x] Add `PULL_REQUEST_TEMPLATE.md`.
- [x] Add [`CODEOWNERS` file](https://github.com/blog/2392-introducing-code-owners) (see: #382).
- [ ] Update `CONTRIBUTING.md` with more information.
- [ ] Add [`SUPPORT` file](https://github.com/blog/2400-support-file-support).
|
1.0
|
Provide more contribution related details - - [x] Add `ISSUE_TEMPLATE.md`.
- [x] Add `PULL_REQUEST_TEMPLATE.md`.
- [x] Add [`CODEOWNERS` file](https://github.com/blog/2392-introducing-code-owners) (see: #382).
- [ ] Update `CONTRIBUTING.md` with more information.
- [ ] Add [`SUPPORT` file](https://github.com/blog/2400-support-file-support).
|
non_process
|
provide more contribution related details add issue template md add pull request template md add see update contributing md with more information add
| 0
|
106,507
| 23,243,713,552
|
IssuesEvent
|
2022-08-03 17:57:23
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
go-to-protobuf: type FooSpec has both field and method named Size
|
kind/bug sig/api-machinery area/code-generation lifecycle/rotten
|
**What happened**:
go-to-protobuf generates `Size()` func which breaks the compilation when the type has `Size` field.
**What you expected to happen**:
go-to-protobuf generation to do not break the compilation.
**How to reproduce it (as minimally and precisely as possible)**:
1. Assume that your type has a field named `Size`. I simple add this type to the `k8s.io/sample-apiserver` types.go.
```go
type FooSpec struct {
Size string `json:"size" protobuf:"bytes,1,opt,name=size"`
}
```
2. Run go-to-protobuf
```
$ go-to-protobuf \
--output-base="${GOPATH}/src" \
--apimachinery-packages='-k8s.io/apimachinery/pkg/util/intstr,-k8s.io/apimachinery/pkg/api/resource,-k8s.io/apimachinery/pkg/runtime/schema,-k8s.io/apimachinery/pkg/runtime,-k8s.io/apimachinery/pkg/apis/meta/v1,-k8s.io/apimachinery/pkg/apis/meta/v1beta1,-k8s.io/api/core/v1,-k8s.io/api/rbac/v1' \
--go-header-file=./hack/custom-boilerplate.go.txt \
--proto-import=./third_party/protobuf \
--proto-import=./vendor \
--packages="k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1"
```
3. Check generated.pb.go and ensure it contains
```go
// ...
func (m *FooSpec) Size() (n int) {
if m == nil {
return 0
}
var l int
_ = l
l = len(m.Size_)
n += 1 + l + sovGenerated(uint64(l))
return n
}
// ...
```
4. Ensure that the compilation fails
```
$ GOOS=linux go build .
# k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1
pkg/apis/wardle/v1alpha1/generated.pb.go:563:12: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:564:18: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:565:49: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size
pkg/apis/wardle/v1alpha1/generated.pb.go:678:11: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:769:36: this.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:1522:5: m.Size_ undefined (type *FooSpec has no field or method Size_)
```
and note the root cause
```
pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size
```
**Anything else we need to know?**:
k8s.io/code-generator version - v1.16.0
**Environment**:
- Kubernetes version (use `kubectl version`):
- Cloud provider or hardware configuration:
- OS (e.g: `cat /etc/os-release`):
- Kernel (e.g. `uname -a`):
- Install tools:
- Network plugin and version (if this is a network-related bug):
- Others:
|
1.0
|
go-to-protobuf: type FooSpec has both field and method named Size - **What happened**:
go-to-protobuf generates `Size()` func which breaks the compilation when the type has `Size` field.
**What you expected to happen**:
go-to-protobuf generation to do not break the compilation.
**How to reproduce it (as minimally and precisely as possible)**:
1. Assume that your type has a field named `Size`. I simple add this type to the `k8s.io/sample-apiserver` types.go.
```go
type FooSpec struct {
Size string `json:"size" protobuf:"bytes,1,opt,name=size"`
}
```
2. Run go-to-protobuf
```
$ go-to-protobuf \
--output-base="${GOPATH}/src" \
--apimachinery-packages='-k8s.io/apimachinery/pkg/util/intstr,-k8s.io/apimachinery/pkg/api/resource,-k8s.io/apimachinery/pkg/runtime/schema,-k8s.io/apimachinery/pkg/runtime,-k8s.io/apimachinery/pkg/apis/meta/v1,-k8s.io/apimachinery/pkg/apis/meta/v1beta1,-k8s.io/api/core/v1,-k8s.io/api/rbac/v1' \
--go-header-file=./hack/custom-boilerplate.go.txt \
--proto-import=./third_party/protobuf \
--proto-import=./vendor \
--packages="k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1"
```
3. Check generated.pb.go and ensure it contains
```go
// ...
func (m *FooSpec) Size() (n int) {
if m == nil {
return 0
}
var l int
_ = l
l = len(m.Size_)
n += 1 + l + sovGenerated(uint64(l))
return n
}
// ...
```
4. Ensure that the compilation fails
```
$ GOOS=linux go build .
# k8s.io/sample-apiserver/pkg/apis/wardle/v1alpha1
pkg/apis/wardle/v1alpha1/generated.pb.go:563:12: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:564:18: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:565:49: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size
pkg/apis/wardle/v1alpha1/generated.pb.go:678:11: m.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:769:36: this.Size_ undefined (type *FooSpec has no field or method Size_)
pkg/apis/wardle/v1alpha1/generated.pb.go:1522:5: m.Size_ undefined (type *FooSpec has no field or method Size_)
```
and note the root cause
```
pkg/apis/wardle/v1alpha1/generated.pb.go:672:6: type FooSpec has both field and method named Size
```
**Anything else we need to know?**:
k8s.io/code-generator version - v1.16.0
**Environment**:
- Kubernetes version (use `kubectl version`):
- Cloud provider or hardware configuration:
- OS (e.g: `cat /etc/os-release`):
- Kernel (e.g. `uname -a`):
- Install tools:
- Network plugin and version (if this is a network-related bug):
- Others:
|
non_process
|
go to protobuf type foospec has both field and method named size what happened go to protobuf generates size func which breaks the compilation when the type has size field what you expected to happen go to protobuf generation to do not break the compilation how to reproduce it as minimally and precisely as possible assume that your type has a field named size i simple add this type to the io sample apiserver types go go type foospec struct size string json size protobuf bytes opt name size run go to protobuf go to protobuf output base gopath src apimachinery packages io apimachinery pkg util intstr io apimachinery pkg api resource io apimachinery pkg runtime schema io apimachinery pkg runtime io apimachinery pkg apis meta io apimachinery pkg apis meta io api core io api rbac go header file hack custom boilerplate go txt proto import third party protobuf proto import vendor packages io sample apiserver pkg apis wardle check generated pb go and ensure it contains go func m foospec size n int if m nil return var l int l l len m size n l sovgenerated l return n ensure that the compilation fails goos linux go build io sample apiserver pkg apis wardle pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go type foospec has both field and method named size pkg apis wardle generated pb go m size undefined type foospec has no field or method size pkg apis wardle generated pb go this size undefined type foospec has no field or method size pkg apis wardle generated pb go m size undefined type foospec has no field or method size and note the root cause pkg apis wardle generated pb go type foospec has both field and method named size anything else we need to know io code generator version environment kubernetes version use kubectl version cloud provider or hardware configuration os e g cat etc os release kernel e g uname a install tools network plugin and version if this is a network related bug others
| 0
|
10,977
| 13,781,068,315
|
IssuesEvent
|
2020-10-08 15:41:47
|
w3c/webauthn
|
https://api.github.com/repos/w3c/webauthn
|
opened
|
Update IANA Registry with Apple attestation format
|
type:process
|
The [IANA Registry](https://www.iana.org/assignments/webauthn/webauthn.xhtml) should be updated to include Apple's new attestation format identifier.
|
1.0
|
Update IANA Registry with Apple attestation format - The [IANA Registry](https://www.iana.org/assignments/webauthn/webauthn.xhtml) should be updated to include Apple's new attestation format identifier.
|
process
|
update iana registry with apple attestation format the should be updated to include apple s new attestation format identifier
| 1
|
11,147
| 13,957,692,968
|
IssuesEvent
|
2020-10-24 08:10:50
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
MT - MITA: Harvesting MT
|
Geoportal Harvesting process MT - Malta
|
Dear Angelo,
Kindly can you perform a harvest for Malta CSW please? as we need to check some changes.
Regards,
Rene
|
1.0
|
MT - MITA: Harvesting MT - Dear Angelo,
Kindly can you perform a harvest for Malta CSW please? as we need to check some changes.
Regards,
Rene
|
process
|
mt mita harvesting mt dear angelo kindly can you perform a harvest for malta csw please as we need to check some changes regards rene
| 1
|
278,480
| 30,702,337,984
|
IssuesEvent
|
2023-07-27 01:21:43
|
nidhi7598/linux-3.0.35
|
https://api.github.com/repos/nidhi7598/linux-3.0.35
|
closed
|
CVE-2021-28972 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2021-28972 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In drivers/pci/hotplug/rpadlpar_sysfs.c in the Linux kernel through 5.11.8, the RPA PCI Hotplug driver has a user-tolerable buffer overflow when writing a new device name to the driver from userspace, allowing userspace to write data to the kernel stack frame directly. This occurs because add_slot_store and remove_slot_store mishandle drc_name '\0' termination, aka CID-cc7a0bb058b8.
<p>Publish Date: 2021-03-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-28972>CVE-2021-28972</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972</a></p>
<p>Release Date: 2021-03-22</p>
<p>Fix Resolution: v4.4.263, v4.9.263, v4.14.227, v4.19.183, v5.4.108, v5.10.26, v5.11.9, v5.12-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-28972 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2021-28972 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/pci/hotplug/rpadlpar_sysfs.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In drivers/pci/hotplug/rpadlpar_sysfs.c in the Linux kernel through 5.11.8, the RPA PCI Hotplug driver has a user-tolerable buffer overflow when writing a new device name to the driver from userspace, allowing userspace to write data to the kernel stack frame directly. This occurs because add_slot_store and remove_slot_store mishandle drc_name '\0' termination, aka CID-cc7a0bb058b8.
<p>Publish Date: 2021-03-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-28972>CVE-2021-28972</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28972</a></p>
<p>Release Date: 2021-03-22</p>
<p>Fix Resolution: v4.4.263, v4.9.263, v4.14.227, v4.19.183, v5.4.108, v5.10.26, v5.11.9, v5.12-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers pci hotplug rpadlpar sysfs c drivers pci hotplug rpadlpar sysfs c drivers pci hotplug rpadlpar sysfs c vulnerability details in drivers pci hotplug rpadlpar sysfs c in the linux kernel through the rpa pci hotplug driver has a user tolerable buffer overflow when writing a new device name to the driver from userspace allowing userspace to write data to the kernel stack frame directly this occurs because add slot store and remove slot store mishandle drc name termination aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
34
| 2,499,975,319
|
IssuesEvent
|
2015-01-08 08:43:15
|
sysown/proxysql-0.2
|
https://api.github.com/repos/sysown/proxysql-0.2
|
opened
|
Hang on LOAD MYSQL QUERY RULES FROM MEMORY
|
ADMIN bug QUERY PROCESSOR
|
```
mysql> select * from mysql_query_rules;
Empty set (0.00 sec)
mysql> show create table mysql_query_rules\G
*************************** 1. row ***************************
table: mysql_query_rules
Create Table: CREATE TABLE mysql_query_rules (rule_id INTEGER PRIMARY KEY AUTOINCREMENT NOT NULL, active INT CHECK (active IN (0,1)) NOT NULL DEFAULT 0, username VARCHAR, schemaname VARCHAR, flagIN INT NOT NULL DEFAULT 0, match_pattern VARCHAR, negate_match_pattern INT CHECK (negate_match_pattern IN (0,1)) NOT NULL DEFAULT 0, flagOUT INT, replace_pattern VARCHAR, destination_hostgroup INT DEFAULT NULL, cache_ttl INT CHECK(cache_ttl > 0), apply INT CHECK(apply IN (0,1)) NOT NULL DEFAULT 0, FOREIGN KEY (destination_hostgroup) REFERENCES mysql_hostgroups (hostgroup_id))
1 row in set (0.00 sec)
mysql> insert into mysql_query_rules values (1, 1, NULL, NULL, 0, '^SELECT.*', 0, 0, NULL, 0, 10, 1);
Query OK, 1 row affected (0.00 sec)
mysql> save mysql query rules to disk;
Query OK, 0 rows affected (0.01 sec)
mysql> load mysql query rules from memory;
```
|
1.0
|
Hang on LOAD MYSQL QUERY RULES FROM MEMORY - ```
mysql> select * from mysql_query_rules;
Empty set (0.00 sec)
mysql> show create table mysql_query_rules\G
*************************** 1. row ***************************
table: mysql_query_rules
Create Table: CREATE TABLE mysql_query_rules (rule_id INTEGER PRIMARY KEY AUTOINCREMENT NOT NULL, active INT CHECK (active IN (0,1)) NOT NULL DEFAULT 0, username VARCHAR, schemaname VARCHAR, flagIN INT NOT NULL DEFAULT 0, match_pattern VARCHAR, negate_match_pattern INT CHECK (negate_match_pattern IN (0,1)) NOT NULL DEFAULT 0, flagOUT INT, replace_pattern VARCHAR, destination_hostgroup INT DEFAULT NULL, cache_ttl INT CHECK(cache_ttl > 0), apply INT CHECK(apply IN (0,1)) NOT NULL DEFAULT 0, FOREIGN KEY (destination_hostgroup) REFERENCES mysql_hostgroups (hostgroup_id))
1 row in set (0.00 sec)
mysql> insert into mysql_query_rules values (1, 1, NULL, NULL, 0, '^SELECT.*', 0, 0, NULL, 0, 10, 1);
Query OK, 1 row affected (0.00 sec)
mysql> save mysql query rules to disk;
Query OK, 0 rows affected (0.01 sec)
mysql> load mysql query rules from memory;
```
|
process
|
hang on load mysql query rules from memory mysql select from mysql query rules empty set sec mysql show create table mysql query rules g row table mysql query rules create table create table mysql query rules rule id integer primary key autoincrement not null active int check active in not null default username varchar schemaname varchar flagin int not null default match pattern varchar negate match pattern int check negate match pattern in not null default flagout int replace pattern varchar destination hostgroup int default null cache ttl int check cache ttl apply int check apply in not null default foreign key destination hostgroup references mysql hostgroups hostgroup id row in set sec mysql insert into mysql query rules values null null select null query ok row affected sec mysql save mysql query rules to disk query ok rows affected sec mysql load mysql query rules from memory
| 1
|
190,350
| 6,817,708,068
|
IssuesEvent
|
2017-11-07 00:50:58
|
FRCteam4909/TGA-Server-2018
|
https://api.github.com/repos/FRCteam4909/TGA-Server-2018
|
opened
|
Analysis Portal & BT Worker Integration (Server->Tablet)
|
Priority: High Status: Assigned Type: Enhancement
|
The analysis portal should be based on Node, and Mongo as per #2.
This should connect with the BT worker via something similar to web hooks, both ways.
Tablet->Server has since been implemented in #4.
|
1.0
|
Analysis Portal & BT Worker Integration (Server->Tablet) - The analysis portal should be based on Node, and Mongo as per #2.
This should connect with the BT worker via something similar to web hooks, both ways.
Tablet->Server has since been implemented in #4.
|
non_process
|
analysis portal bt worker integration server tablet the analysis portal should be based on node and mongo as per this should connect with the bt worker via something similar to web hooks both ways tablet server has since been implemented in
| 0
|
591,007
| 17,792,932,822
|
IssuesEvent
|
2021-08-31 18:24:00
|
status-im/status-desktop
|
https://api.github.com/repos/status-im/status-desktop
|
closed
|
no identicon/profile picture shown and wrong timestamp format in search results
|
bug ui priority F2: important
|
Search results should display the identicon or profile picture if available. Also timestamp is shown as a number instead of correct date format.

|
1.0
|
no identicon/profile picture shown and wrong timestamp format in search results - Search results should display the identicon or profile picture if available. Also timestamp is shown as a number instead of correct date format.

|
non_process
|
no identicon profile picture shown and wrong timestamp format in search results search results should display the identicon or profile picture if available also timestamp is shown as a number instead of correct date format
| 0
|
20,225
| 3,799,106,924
|
IssuesEvent
|
2016-03-23 14:57:01
|
ensime/scala-debugger
|
https://api.github.com/repos/ensime/scala-debugger
|
closed
|
Create Drone cache image for testing
|
enhancement for test Needs Review
|
Currently, the Drone setup downloads the sbt plugins (including Coursier) as well as all dependencies on each run. It would be a very good idea to produce a custom Docker image that contains these dependencies to both speed up run time and avoid hammering Maven Central (and other repositories).
I'd love for the auto-commit caching to work (it looks like Drone is doing something to do with committing a new image), but for now this seems to be the best approach. This is what you were talking about earlier, right, @fommil?
|
1.0
|
Create Drone cache image for testing - Currently, the Drone setup downloads the sbt plugins (including Coursier) as well as all dependencies on each run. It would be a very good idea to produce a custom Docker image that contains these dependencies to both speed up run time and avoid hammering Maven Central (and other repositories).
I'd love for the auto-commit caching to work (it looks like Drone is doing something to do with committing a new image), but for now this seems to be the best approach. This is what you were talking about earlier, right, @fommil?
|
non_process
|
create drone cache image for testing currently the drone setup downloads the sbt plugins including coursier as well as all dependencies on each run it would be a very good idea to produce a custom docker image that contains these dependencies to both speed up run time and avoid hammering maven central and other repositories i d love for the auto commit caching to work it looks like drone is doing something to do with committing a new image but for now this seems to be the best approach this is what you were talking about earlier right fommil
| 0
|
13,590
| 16,162,951,516
|
IssuesEvent
|
2021-05-01 01:27:17
|
tdwg/chrono
|
https://api.github.com/repos/tdwg/chrono
|
closed
|
Handling of IRI values
|
Process - prepare for Executive review Question - answered
|
Is there a mechanism to supply IRI values for chrono:chronometricAgeDeterminedBy, either with an id term chrono:chronometricAgeDeterminedByID or an IRI namespace chronoiri:chronometricAgeDeterminedBy? Such a mechanism would be needed to allow an RDF representation to express a list of chronometric age determiners as repeated id terms, each pointing to a determiner agent by ORCID or other agent identifier.
_Originally posted by @chicoreus in https://github.com/tdwg/chrono/issues/15#issuecomment-732198397_
|
1.0
|
Handling of IRI values - Is there a mechanism to supply IRI values for chrono:chronometricAgeDeterminedBy, either with an id term chrono:chronometricAgeDeterminedByID or an IRI namespace chronoiri:chronometricAgeDeterminedBy? Such a mechanism would be needed to allow an RDF representation to express a list of chronometric age determiners as repeated id terms, each pointing to a determiner agent by ORCID or other agent identifier.
_Originally posted by @chicoreus in https://github.com/tdwg/chrono/issues/15#issuecomment-732198397_
|
process
|
handling of iri values is there a mechanism to supply iri values for chrono chronometricagedeterminedby either with an id term chrono chronometricagedeterminedbyid or an iri namespace chronoiri chronometricagedeterminedby such a mechanism would be needed to allow an rdf representation to express a list of chronometric age determiners as repeated id terms each pointing to a determiner agent by orcid or other agent identifier originally posted by chicoreus in
| 1
|
21,656
| 30,105,856,141
|
IssuesEvent
|
2023-06-30 01:07:53
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
hpcflow-new2 0.2.0a55 has 1 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/hpcflow-new2
https://inspector.pypi.io/project/hpcflow-new2
```{
"dependency": "hpcflow-new2",
"version": "0.2.0a55",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "hpcflow_new2-0.2.0a55/hpcflow/sdk/helper/helper.py:118",
"code": " proc = subprocess.Popen(\n args=args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n **kwargs,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpbko46hx_/hpcflow-new2"
}
}```
|
1.0
|
hpcflow-new2 0.2.0a55 has 1 GuardDog issues - https://pypi.org/project/hpcflow-new2
https://inspector.pypi.io/project/hpcflow-new2
```{
"dependency": "hpcflow-new2",
"version": "0.2.0a55",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "hpcflow_new2-0.2.0a55/hpcflow/sdk/helper/helper.py:118",
"code": " proc = subprocess.Popen(\n args=args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n **kwargs,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpbko46hx_/hpcflow-new2"
}
}```
|
process
|
hpcflow has guarddog issues dependency hpcflow version result issues errors results silent process execution location hpcflow hpcflow sdk helper helper py code proc subprocess popen n args args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n kwargs n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp hpcflow
| 1
|
790,943
| 27,843,833,896
|
IssuesEvent
|
2023-03-20 14:22:56
|
GSM-MSG/GCMS-FrontEnd-V2
|
https://api.github.com/repos/GSM-MSG/GCMS-FrontEnd-V2
|
closed
|
동아리 디테일에서 설명이나 제목이 space나 \n을 안 읽어요
|
2️⃣ Priority: Medium ♻️ Refactor 🐞 Bug
|
### Describe
<img width="1392" alt="스크린샷 2023-03-20 오전 1 15 55" src="https://user-images.githubusercontent.com/81547954/226189371-f2c619ef-847b-4ba7-a2aa-fe3f122fe558.png">
보세요!
### Additional
_No response_
|
1.0
|
동아리 디테일에서 설명이나 제목이 space나 \n을 안 읽어요 - ### Describe
<img width="1392" alt="스크린샷 2023-03-20 오전 1 15 55" src="https://user-images.githubusercontent.com/81547954/226189371-f2c619ef-847b-4ba7-a2aa-fe3f122fe558.png">
보세요!
### Additional
_No response_
|
non_process
|
동아리 디테일에서 설명이나 제목이 space나 n을 안 읽어요 describe img width alt 스크린샷 오전 src 보세요 additional no response
| 0
|
14,304
| 17,290,977,235
|
IssuesEvent
|
2021-07-24 18:46:56
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
@parcel/transformer-sass: Fails to resolve path to node module
|
:bug: Bug CSS Preprocessing ✨ Parcel 2
|
<!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
# 🐛 bug report
In the newly released parcel 2.0.0-beta.2 (and nightly builds before that) transformer-sass can’t resolve paths to installed node modules.
The error looks similar to https://github.com/parcel-bundler/parcel/issues/5813 but happens on build and serve.
## 🎛 Configuration (.babelrc, package.json, cli command)
<!--- If describing a bug, tell us what your babel configuration looks like -->
package.json
```js
"dependencies": {
"autoprefixer": "~10.2.5",
"core-js": "~3.9.1",
"foundation-sites": "6.6.3",
"husky": "~5.1.3",
"jquery": "~3.6.0",
"motion-ui": "~2.0.3",
"parcel": "2.0.0-beta.2",
"postcss": "~8.2.7",
"sass": "~1.32.8",
},
// automatically added by parcel:
"devDependencies": {
"@babel/core": "^7.13.10",
"@parcel/transformer-sass": "2.0.0-beta.2"
}
```
.scssrc
```js
{
"includePaths": ["node_modules"]
}
```
In my scss files, I'm importing scss files from foundation-sites, e.g.
```scss
@import "foundation-sites/scss/global";
```
## 🤔 Expected Behavior
I expect the scss file to be included (used to still work with parcel 2.0.0-beta.1).
## 😯 Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
`yarn parcel serve` and `yarn parcel build` throw an error. No css file is build.
<!--- If you are seeing an error, please include the full error message and stack trace -->
```
Build failed.
@parcel/transformer-sass: Got unexpected null
╷
19 │ @import "foundation-sites/scss/global";
│ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
╵
scss/main.scss 19:9 root stylesheet
Error: Got unexpected null
╷
19 │ @import "foundation-sites/scss/global";
│ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
╵
scss/main.scss 19:9 root stylesheet
at Object._newRenderError (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13537:19)
at Object._wrapException (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13374:16)
at _render_closure1.call$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:80373:21)
at _RootZone.runBinary$3$3 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:27269:18)
at _FutureListener.handleError$1 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25797:19)
at _Future__propagateToListeners_handleError.call$0 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:26094:49)
at Object._Future__propagateToListeners (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4543:77)
at _Future._completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25927:9)
at _AsyncAwaitCompleter.completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25270:12)
at Object._asyncRethrow (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4292:17)
```
On previous tries the error message was "Failed to resolve" with an identical stack trace.
## 💁 Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug -->
Maybe scss files need to be imported differently now?
## 🔦 Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
The issue makes any recent version of parcel 2 unusable for us. We need to be able to import scss files from other node modules.
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## 💻 Code Sample
<!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue -->
## 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 2.0.0-beta.2
| Node | 12.19.0
| npm/Yarn | yarn 1.22.5
| Operating System | Ubuntu 20.10
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
|
1.0
|
@parcel/transformer-sass: Fails to resolve path to node module - <!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
# 🐛 bug report
In the newly released parcel 2.0.0-beta.2 (and nightly builds before that) transformer-sass can’t resolve paths to installed node modules.
The error looks similar to https://github.com/parcel-bundler/parcel/issues/5813 but happens on build and serve.
## 🎛 Configuration (.babelrc, package.json, cli command)
<!--- If describing a bug, tell us what your babel configuration looks like -->
package.json
```js
"dependencies": {
"autoprefixer": "~10.2.5",
"core-js": "~3.9.1",
"foundation-sites": "6.6.3",
"husky": "~5.1.3",
"jquery": "~3.6.0",
"motion-ui": "~2.0.3",
"parcel": "2.0.0-beta.2",
"postcss": "~8.2.7",
"sass": "~1.32.8",
},
// automatically added by parcel:
"devDependencies": {
"@babel/core": "^7.13.10",
"@parcel/transformer-sass": "2.0.0-beta.2"
}
```
.scssrc
```js
{
"includePaths": ["node_modules"]
}
```
In my scss files, I'm importing scss files from foundation-sites, e.g.
```scss
@import "foundation-sites/scss/global";
```
## 🤔 Expected Behavior
I expect the scss file to be included (used to still work with parcel 2.0.0-beta.1).
## 😯 Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
`yarn parcel serve` and `yarn parcel build` throw an error. No css file is build.
<!--- If you are seeing an error, please include the full error message and stack trace -->
```
Build failed.
@parcel/transformer-sass: Got unexpected null
╷
19 │ @import "foundation-sites/scss/global";
│ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
╵
scss/main.scss 19:9 root stylesheet
Error: Got unexpected null
╷
19 │ @import "foundation-sites/scss/global";
│ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
╵
scss/main.scss 19:9 root stylesheet
at Object._newRenderError (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13537:19)
at Object._wrapException (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:13374:16)
at _render_closure1.call$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:80373:21)
at _RootZone.runBinary$3$3 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:27269:18)
at _FutureListener.handleError$1 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25797:19)
at _Future__propagateToListeners_handleError.call$0 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:26094:49)
at Object._Future__propagateToListeners (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4543:77)
at _Future._completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25927:9)
at _AsyncAwaitCompleter.completeError$2 (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:25270:12)
at Object._asyncRethrow (/var/code/themes/mo-foundation-base/node_modules/sass/sass.dart.js:4292:17)
```
On previous tries the error message was "Failed to resolve" with an identical stack trace.
## 💁 Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug -->
Maybe scss files need to be imported differently now?
## 🔦 Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
The issue makes any recent version of parcel 2 unusable for us. We need to be able to import scss files from other node modules.
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## 💻 Code Sample
<!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue -->
## 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 2.0.0-beta.2
| Node | 12.19.0
| npm/Yarn | yarn 1.22.5
| Operating System | Ubuntu 20.10
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
|
process
|
parcel transformer sass fails to resolve path to node module thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before 🐛 bug report in the newly released parcel beta and nightly builds before that transformer sass can’t resolve paths to installed node modules the error looks similar to but happens on build and serve 🎛 configuration babelrc package json cli command package json js dependencies autoprefixer core js foundation sites husky jquery motion ui parcel beta postcss sass automatically added by parcel devdependencies babel core parcel transformer sass beta scssrc js includepaths in my scss files i m importing scss files from foundation sites e g scss import foundation sites scss global 🤔 expected behavior i expect the scss file to be included used to still work with parcel beta 😯 current behavior yarn parcel serve and yarn parcel build throw an error no css file is build build failed parcel transformer sass got unexpected null ╷ │ import foundation sites scss global │ ╵ scss main scss root stylesheet error got unexpected null ╷ │ import foundation sites scss global │ ╵ scss main scss root stylesheet at object newrendererror var code themes mo foundation base node modules sass sass dart js at object wrapexception var code themes mo foundation base node modules sass sass dart js at render call var code themes mo foundation base node modules sass sass dart js at rootzone runbinary var code themes mo foundation base node modules sass sass dart js at futurelistener handleerror var code themes mo foundation base node modules sass sass dart js at future propagatetolisteners handleerror call var code themes mo foundation base node modules sass sass dart js at object future propagatetolisteners var code themes mo foundation base node modules sass sass dart js at future completeerror var code themes mo foundation base node modules sass sass dart js at asyncawaitcompleter completeerror var code themes mo foundation base node modules sass sass dart js at object asyncrethrow var code themes mo foundation base node modules sass sass dart js on previous tries the error message was failed to resolve with an identical stack trace 💁 possible solution maybe scss files need to be imported differently now 🔦 context the issue makes any recent version of parcel unusable for us we need to be able to import scss files from other node modules 💻 code sample 🌍 your environment software version s parcel beta node npm yarn yarn operating system ubuntu love parcel please consider supporting our collective 👉
| 1
|
55,860
| 3,075,045,331
|
IssuesEvent
|
2015-08-20 11:16:02
|
cdnjs/cdnjs
|
https://api.github.com/repos/cdnjs/cdnjs
|
closed
|
[bug]npm auto-update can not pick every versions properly at once
|
Bug - High Priority Help wanted
|
Need to run the auto-update process for many times to pick all the versions
> error downloading `123` of `xxx` it didnt exist: [object Object]null
|
1.0
|
[bug]npm auto-update can not pick every versions properly at once - Need to run the auto-update process for many times to pick all the versions
> error downloading `123` of `xxx` it didnt exist: [object Object]null
|
non_process
|
npm auto update can not pick every versions properly at once need to run the auto update process for many times to pick all the versions error downloading of xxx it didnt exist null
| 0
|
7,757
| 10,878,344,777
|
IssuesEvent
|
2019-11-16 17:00:52
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
closed
|
Refactor workflow in App-Frontend
|
area/process area/receipt breaking-change kind/user-story solution/app-frontend team/tamagotchi
|
## Description
Based on the analysis in #13678, refactor Workflowstep / Formfiller / Receipt
## BREAKING CHANGE
Note that this issue introduces breaking changes for existing applications. How to fix is documented [on altinn docs.](https://docs.altinn.studio/known-issues/breaking-changes/#breaking-change-data-element-validation-returns-an-internal-server-error-to-app-frontend)
## Desired process flow

## Considerations
Input (beyond tasks) on how the user story should be solved can be put here.
## Acceptance criteria
- App frontend are updated to use new process api
- The 'Send inn' step is removed
- All other functionality works as before
## Out of scope
- Generic process handling is scoped out
## Tasks
- [x] Verify that this issue meets DoP (remove unused text, add missing text/parameters/labels, verify tasks).
### Specification tasks
- [x] Test design / decide test need
### Development tasks
- [x] Refactor the "workflow resource" to `shared/resources/process`
- [x] Move src\react-apps\applications\runtime\src\features\form\workflow
- [x] Refactor src\react-apps\applications\runtime\src\containers\WorkflowStep.tsx
- [x] Rename "set current state" action to "write current workflow state to store" or similar... *needs to be discussed*
- [x] Actions,Sagas,Reducers from src\react-apps\applications\runtime\src\features\form\workflow\ needs to be moved to new folder with new structure.
- [x] Create `getProcessStateSaga` that gets current process state. Place under `shared/resources/process/getProcessState`
- [x] Update how we handle "Update" and "Complete" actions to the controller @lorang92
- [x] Update should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/data/{dataGuid}`
- [x] Complete should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/process/completeProcess`
- [x] Create shared ProcessStep enum
- [x] Rename Workflow containers/components to Process
### QA
- [x] Documentation (if relevant)
- [x] Manual test (if needed)
- [x] Automated test (if needed)
|
1.0
|
Refactor workflow in App-Frontend - ## Description
Based on the analysis in #13678, refactor Workflowstep / Formfiller / Receipt
## BREAKING CHANGE
Note that this issue introduces breaking changes for existing applications. How to fix is documented [on altinn docs.](https://docs.altinn.studio/known-issues/breaking-changes/#breaking-change-data-element-validation-returns-an-internal-server-error-to-app-frontend)
## Desired process flow

## Considerations
Input (beyond tasks) on how the user story should be solved can be put here.
## Acceptance criteria
- App frontend are updated to use new process api
- The 'Send inn' step is removed
- All other functionality works as before
## Out of scope
- Generic process handling is scoped out
## Tasks
- [x] Verify that this issue meets DoP (remove unused text, add missing text/parameters/labels, verify tasks).
### Specification tasks
- [x] Test design / decide test need
### Development tasks
- [x] Refactor the "workflow resource" to `shared/resources/process`
- [x] Move src\react-apps\applications\runtime\src\features\form\workflow
- [x] Refactor src\react-apps\applications\runtime\src\containers\WorkflowStep.tsx
- [x] Rename "set current state" action to "write current workflow state to store" or similar... *needs to be discussed*
- [x] Actions,Sagas,Reducers from src\react-apps\applications\runtime\src\features\form\workflow\ needs to be moved to new folder with new structure.
- [x] Create `getProcessStateSaga` that gets current process state. Place under `shared/resources/process/getProcessState`
- [x] Update how we handle "Update" and "Complete" actions to the controller @lorang92
- [x] Update should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/data/{dataGuid}`
- [x] Complete should call `/{org}/{app}/instances/{instanceOwnerId}/{instanceGuid}/process/completeProcess`
- [x] Create shared ProcessStep enum
- [x] Rename Workflow containers/components to Process
### QA
- [x] Documentation (if relevant)
- [x] Manual test (if needed)
- [x] Automated test (if needed)
|
process
|
refactor workflow in app frontend description based on the analysis in refactor workflowstep formfiller receipt breaking change note that this issue introduces breaking changes for existing applications how to fix is documented desired process flow considerations input beyond tasks on how the user story should be solved can be put here acceptance criteria app frontend are updated to use new process api the send inn step is removed all other functionality works as before out of scope generic process handling is scoped out tasks verify that this issue meets dop remove unused text add missing text parameters labels verify tasks specification tasks test design decide test need development tasks refactor the workflow resource to shared resources process move src react apps applications runtime src features form workflow refactor src react apps applications runtime src containers workflowstep tsx rename set current state action to write current workflow state to store or similar needs to be discussed actions sagas reducers from src react apps applications runtime src features form workflow needs to be moved to new folder with new structure create getprocessstatesaga that gets current process state place under shared resources process getprocessstate update how we handle update and complete actions to the controller update should call org app instances instanceownerid instanceguid data dataguid complete should call org app instances instanceownerid instanceguid process completeprocess create shared processstep enum rename workflow containers components to process qa documentation if relevant manual test if needed automated test if needed
| 1
|
822,994
| 30,923,281,124
|
IssuesEvent
|
2023-08-06 06:56:38
|
unitystation/unitystation
|
https://api.github.com/repos/unitystation/unitystation
|
closed
|
The text size in the player list is massive
|
Type: Feature Priority: Low Good First Issue
|
Thanks to the text size, the player list can only fit something like five people at the moment. We might wanna scale that down just a tad.
https://i.imgur.com/un9DBVd.png
|
1.0
|
The text size in the player list is massive - Thanks to the text size, the player list can only fit something like five people at the moment. We might wanna scale that down just a tad.
https://i.imgur.com/un9DBVd.png
|
non_process
|
the text size in the player list is massive thanks to the text size the player list can only fit something like five people at the moment we might wanna scale that down just a tad
| 0
|
399,358
| 11,747,581,358
|
IssuesEvent
|
2020-03-12 13:52:37
|
AbsaOSS/hyperdrive-trigger
|
https://api.github.com/repos/AbsaOSS/hyperdrive-trigger
|
opened
|
Tables - use identity instead serial in table implementations
|
backend bug priority: medium
|
Instead of serial, you should probably use identity, since identity is an SQL-Standard whereas serial is postgres specific. E.g. "id" BIGINT GENERATED BY DEFAULT AS IDENTITY NOT NULL PRIMARY KEY
|
1.0
|
Tables - use identity instead serial in table implementations - Instead of serial, you should probably use identity, since identity is an SQL-Standard whereas serial is postgres specific. E.g. "id" BIGINT GENERATED BY DEFAULT AS IDENTITY NOT NULL PRIMARY KEY
|
non_process
|
tables use identity instead serial in table implementations instead of serial you should probably use identity since identity is an sql standard whereas serial is postgres specific e g id bigint generated by default as identity not null primary key
| 0
|
21,549
| 29,865,410,710
|
IssuesEvent
|
2023-06-20 03:04:17
|
cncf/tag-security
|
https://api.github.com/repos/cncf/tag-security
|
closed
|
[Sec Assess WG] Mapping Security Assessments to TOC Process
|
help wanted assessment-process suggestion inactive
|
This issue was created from results of the Security Assessment Improvement Working Group (https://github.com/cncf/sig-security/issues/167#issuecomment-714514142).
# Mapping Security Assessments to TOC Process
## Premise
- The security assessment aims to tie into the CNCF project process, but it is not clear how
- There is no current agreed upon requirement for the process for what is asked of a SIG
## Ideas
- better document the Due Diligence process and document
- Need more detail on when the SIG recommends for the next CNCF phase
- not tie it too directly to TOC process
- CNCF needs to explicitly define the requirement for projects at each level to go through assessment
- Map security assessment process to process of CNCF
- Are assessments necessary/mandated by the TOC to move from one stage to another ? We need to make projects realize it's in their interest to invest time in going through the assessment process.
## Action Items
- [ ] Create internal mapping of TOC mapping for SIG-Security (@lumjjb, @itaysk)
- [ ] Propose a hard requirement for TOC process (looking at frequency/data of incubating/graduating projects)
## Logistics
- [ ] Contributors (For multiple contributors, 1 lead to coordinate)
- [ ] SIG-Representative
|
1.0
|
[Sec Assess WG] Mapping Security Assessments to TOC Process - This issue was created from results of the Security Assessment Improvement Working Group (https://github.com/cncf/sig-security/issues/167#issuecomment-714514142).
# Mapping Security Assessments to TOC Process
## Premise
- The security assessment aims to tie into the CNCF project process, but it is not clear how
- There is no current agreed upon requirement for the process for what is asked of a SIG
## Ideas
- better document the Due Diligence process and document
- Need more detail on when the SIG recommends for the next CNCF phase
- not tie it too directly to TOC process
- CNCF needs to explicitly define the requirement for projects at each level to go through assessment
- Map security assessment process to process of CNCF
- Are assessments necessary/mandated by the TOC to move from one stage to another ? We need to make projects realize it's in their interest to invest time in going through the assessment process.
## Action Items
- [ ] Create internal mapping of TOC mapping for SIG-Security (@lumjjb, @itaysk)
- [ ] Propose a hard requirement for TOC process (looking at frequency/data of incubating/graduating projects)
## Logistics
- [ ] Contributors (For multiple contributors, 1 lead to coordinate)
- [ ] SIG-Representative
|
process
|
mapping security assessments to toc process this issue was created from results of the security assessment improvement working group mapping security assessments to toc process premise the security assessment aims to tie into the cncf project process but it is not clear how there is no current agreed upon requirement for the process for what is asked of a sig ideas better document the due diligence process and document need more detail on when the sig recommends for the next cncf phase not tie it too directly to toc process cncf needs to explicitly define the requirement for projects at each level to go through assessment map security assessment process to process of cncf are assessments necessary mandated by the toc to move from one stage to another we need to make projects realize it s in their interest to invest time in going through the assessment process action items create internal mapping of toc mapping for sig security lumjjb itaysk propose a hard requirement for toc process looking at frequency data of incubating graduating projects logistics contributors for multiple contributors lead to coordinate sig representative
| 1
|
15,576
| 19,703,508,290
|
IssuesEvent
|
2022-01-12 19:08:19
|
googleapis/nodejs-filestore
|
https://api.github.com/repos/googleapis/nodejs-filestore
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'filestore' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'filestore' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname filestore invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
14,714
| 17,926,562,100
|
IssuesEvent
|
2021-09-10 00:51:39
|
Leviatan-Analytics/LA-data-processing
|
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
|
closed
|
Implement text recognition and ward detection scripts parallelization [3]
|
Data Processing Week 2 Sprint 4
|
Use the knowledge of the previous research and implement the parallelization of the scripts.
|
1.0
|
Implement text recognition and ward detection scripts parallelization [3] - Use the knowledge of the previous research and implement the parallelization of the scripts.
|
process
|
implement text recognition and ward detection scripts parallelization use the knowledge of the previous research and implement the parallelization of the scripts
| 1
|
55,837
| 11,470,803,925
|
IssuesEvent
|
2020-02-09 06:27:18
|
eclipse-theia/theia
|
https://api.github.com/repos/eclipse-theia/theia
|
opened
|
plugin tree rows are bogusly rendered
|
tree ui/ux vscode
|
Some recent PR, broke plugin tree rendering, rows are not rendered inline anymore
<img width="1904" alt="Screenshot 2020-02-09 at 07 25 26" src="https://user-images.githubusercontent.com/3082655/74097556-7fdbf100-4b0d-11ea-9c27-f7b8c934fb47.png">
@vince-fugnitto Do you remember which PR? Could we revert it? It is quite bad now.
|
1.0
|
plugin tree rows are bogusly rendered - Some recent PR, broke plugin tree rendering, rows are not rendered inline anymore
<img width="1904" alt="Screenshot 2020-02-09 at 07 25 26" src="https://user-images.githubusercontent.com/3082655/74097556-7fdbf100-4b0d-11ea-9c27-f7b8c934fb47.png">
@vince-fugnitto Do you remember which PR? Could we revert it? It is quite bad now.
|
non_process
|
plugin tree rows are bogusly rendered some recent pr broke plugin tree rendering rows are not rendered inline anymore img width alt screenshot at src vince fugnitto do you remember which pr could we revert it it is quite bad now
| 0
|
14,121
| 10,616,442,453
|
IssuesEvent
|
2019-10-12 11:46:24
|
python-pillow/Pillow
|
https://api.github.com/repos/python-pillow/Pillow
|
closed
|
Automated security fixes with Dependabot?
|
Infrastructure
|
GitHub will be sending automated PRs with security fixes for accounts with security alerts enabled.
Pillow does not have any Python dependencies for production code, only testing, but shall we enabled the setting to see how it goes?
It's at https://github.com/python-pillow/Pillow/network/alerts (I don't have access).
More info:
* https://github.blog/2019-05-23-introducing-new-ways-to-keep-your-code-secure/#automated-security-fixes-with-dependabot
* https://help.github.com/en/articles/configuring-automated-security-fixes
|
1.0
|
Automated security fixes with Dependabot? - GitHub will be sending automated PRs with security fixes for accounts with security alerts enabled.
Pillow does not have any Python dependencies for production code, only testing, but shall we enabled the setting to see how it goes?
It's at https://github.com/python-pillow/Pillow/network/alerts (I don't have access).
More info:
* https://github.blog/2019-05-23-introducing-new-ways-to-keep-your-code-secure/#automated-security-fixes-with-dependabot
* https://help.github.com/en/articles/configuring-automated-security-fixes
|
non_process
|
automated security fixes with dependabot github will be sending automated prs with security fixes for accounts with security alerts enabled pillow does not have any python dependencies for production code only testing but shall we enabled the setting to see how it goes it s at i don t have access more info
| 0
|
5,689
| 8,560,128,212
|
IssuesEvent
|
2018-11-08 23:44:45
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
Upgrade testing
|
area/API area/test-and-release kind/feature kind/process
|
<!--
/area API
/area test-and-release
/kind feature
/kind process
/assign @jonjohnsonjr
-->
We should have testing that verifies that upgrading from our last release to HEAD works.
|
1.0
|
Upgrade testing - <!--
/area API
/area test-and-release
/kind feature
/kind process
/assign @jonjohnsonjr
-->
We should have testing that verifies that upgrading from our last release to HEAD works.
|
process
|
upgrade testing area api area test and release kind feature kind process assign jonjohnsonjr we should have testing that verifies that upgrading from our last release to head works
| 1
|
16,713
| 21,871,329,092
|
IssuesEvent
|
2022-05-19 05:45:10
|
keras-team/keras-cv
|
https://api.github.com/repos/keras-team/keras-cv
|
closed
|
`MaybeApply` high order abstraction
|
contribution-welcome preprocessing
|
We should consider an implementation for a `MaybeApply` layer:
```
maybe_solarize = MaybeApply(Solarization(), rate=0.5)
# maybe_solarize solarizes half of the images passed to it
```
https://github.com/chjort/chambers/blob/b796f507b7353fb141b5c8b026b811a023d303e6/chambers/augmentations/image_augmentations.py#L514
Notes:
- we can rely on this in RandomAugmentationPipeline
- this must be element-wise, not batch wise!
|
1.0
|
`MaybeApply` high order abstraction - We should consider an implementation for a `MaybeApply` layer:
```
maybe_solarize = MaybeApply(Solarization(), rate=0.5)
# maybe_solarize solarizes half of the images passed to it
```
https://github.com/chjort/chambers/blob/b796f507b7353fb141b5c8b026b811a023d303e6/chambers/augmentations/image_augmentations.py#L514
Notes:
- we can rely on this in RandomAugmentationPipeline
- this must be element-wise, not batch wise!
|
process
|
maybeapply high order abstraction we should consider an implementation for a maybeapply layer maybe solarize maybeapply solarization rate maybe solarize solarizes half of the images passed to it notes we can rely on this in randomaugmentationpipeline this must be element wise not batch wise
| 1
|
194,963
| 22,281,615,794
|
IssuesEvent
|
2022-06-11 01:14:31
|
pazhanivel07/linux-4.19.72
|
https://api.github.com/repos/pazhanivel07/linux-4.19.72
|
reopened
|
CVE-2021-20177 (Medium) detected in linux-yoctov5.4.51, linux-yoctov5.4.51
|
security vulnerability
|
## CVE-2021-20177 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linux-yoctov5.4.51</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel's implementation of string matching within a packet. A privileged user (with root or CAP_NET_ADMIN) when inserting iptables rules could insert a rule which can panic the system. Kernel before kernel 5.5-rc1 is affected.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20177>CVE-2021-20177</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-20177">https://www.linuxkernelcves.com/cves/CVE-2021-20177</a></p>
<p>Release Date: 2021-05-26</p>
<p>Fix Resolution: v5.5-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-20177 (Medium) detected in linux-yoctov5.4.51, linux-yoctov5.4.51 - ## CVE-2021-20177 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linux-yoctov5.4.51</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel's implementation of string matching within a packet. A privileged user (with root or CAP_NET_ADMIN) when inserting iptables rules could insert a rule which can panic the system. Kernel before kernel 5.5-rc1 is affected.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20177>CVE-2021-20177</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-20177">https://www.linuxkernelcves.com/cves/CVE-2021-20177</a></p>
<p>Release Date: 2021-05-26</p>
<p>Fix Resolution: v5.5-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux linux cve medium severity vulnerability vulnerable libraries linux linux vulnerability details a flaw was found in the linux kernel s implementation of string matching within a packet a privileged user with root or cap net admin when inserting iptables rules could insert a rule which can panic the system kernel before kernel is affected publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
3,451
| 6,542,244,402
|
IssuesEvent
|
2017-09-02 02:54:51
|
gaocegege/Processing.R
|
https://api.github.com/repos/gaocegege/Processing.R
|
closed
|
Replace ant with maven or gradle
|
community/processing priority/p3 size/no-idea status/to-be-claimed type/feasibility-research
|
Now Processing.R uses maven-ant-tasks to install maven dependencies in ant. If we could use maven instead, it may be better.
|
1.0
|
Replace ant with maven or gradle - Now Processing.R uses maven-ant-tasks to install maven dependencies in ant. If we could use maven instead, it may be better.
|
process
|
replace ant with maven or gradle now processing r uses maven ant tasks to install maven dependencies in ant if we could use maven instead it may be better
| 1
|
19,311
| 25,466,751,008
|
IssuesEvent
|
2022-11-25 05:43:13
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[IDP] [PM] UI Issue when user enters wrong password in the auth server
|
Bug P1 Process: Fixed Process: Tested QA Process: Tested dev Auth server
|
UI Issue when user enters wrong password in the 'AUTH' server
**AR:**

**ER:**

|
3.0
|
[IDP] [PM] UI Issue when user enters wrong password in the auth server - UI Issue when user enters wrong password in the 'AUTH' server
**AR:**

**ER:**

|
process
|
ui issue when user enters wrong password in the auth server ui issue when user enters wrong password in the auth server ar er
| 1
|
367,354
| 25,734,280,477
|
IssuesEvent
|
2022-12-07 23:01:51
|
cncf/cnf-wg
|
https://api.github.com/repos/cncf/cnf-wg
|
opened
|
Suggestion: Update categories names to be the same types of words
|
documentation enhancement
|
Some are properties, some are areas, etc. Suggestion: use words that are of one type for each category name.
Examples:
- [Observability and Diagnostics](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#90-observability-and-diagnostics)
- [Security](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#70-security)
|
1.0
|
Suggestion: Update categories names to be the same types of words - Some are properties, some are areas, etc. Suggestion: use words that are of one type for each category name.
Examples:
- [Observability and Diagnostics](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#90-observability-and-diagnostics)
- [Security](https://github.com/cncf/cnf-wg/blob/main/doc/best_cnf_dev.md#70-security)
|
non_process
|
suggestion update categories names to be the same types of words some are properties some are areas etc suggestion use words that are of one type for each category name examples
| 0
|
2,436
| 5,216,170,327
|
IssuesEvent
|
2017-01-26 09:22:15
|
QCoDeS/Qcodes
|
https://api.github.com/repos/QCoDeS/Qcodes
|
closed
|
PyQtgraph live plot does not update
|
bug mulitprocessing
|
Issue made to get rid of the hack in PR #250 .
### Steps to reproduce
1. Run Tutorial notebook from #250
``` python
data = qc.Loop(c0.sweep(-20,20,0.1), delay=0.003).each(meter.amplitude).run(name='testsweep')
plotQ = qc.QtPlot()
plotQ.add(data.meter_amplitude)
```
### Expected behaviour
Expect to see a live updating plot of the data
### Actual behaviour
Plots only what is in "data" the moment the plot function is called.
### Workaround
``` python
while data.sync():
plotQ.update()
plotQ.update()
```
The following code will keep updating the plot (in the main thread) while the data taking is not finished yet.
### System
**operating system**
OS X El Capitan 10.11.5 (15F34)
**qcodes branch**
"examples" (only contains notebook so should be close to master)
**qcodes commit**
Commit: 007e5f95ac5
|
1.0
|
PyQtgraph live plot does not update - Issue made to get rid of the hack in PR #250 .
### Steps to reproduce
1. Run Tutorial notebook from #250
``` python
data = qc.Loop(c0.sweep(-20,20,0.1), delay=0.003).each(meter.amplitude).run(name='testsweep')
plotQ = qc.QtPlot()
plotQ.add(data.meter_amplitude)
```
### Expected behaviour
Expect to see a live updating plot of the data
### Actual behaviour
Plots only what is in "data" the moment the plot function is called.
### Workaround
``` python
while data.sync():
plotQ.update()
plotQ.update()
```
The following code will keep updating the plot (in the main thread) while the data taking is not finished yet.
### System
**operating system**
OS X El Capitan 10.11.5 (15F34)
**qcodes branch**
"examples" (only contains notebook so should be close to master)
**qcodes commit**
Commit: 007e5f95ac5
|
process
|
pyqtgraph live plot does not update issue made to get rid of the hack in pr steps to reproduce run tutorial notebook from python data qc loop sweep delay each meter amplitude run name testsweep plotq qc qtplot plotq add data meter amplitude expected behaviour expect to see a live updating plot of the data actual behaviour plots only what is in data the moment the plot function is called workaround python while data sync plotq update plotq update the following code will keep updating the plot in the main thread while the data taking is not finished yet system operating system os x el capitan qcodes branch examples only contains notebook so should be close to master qcodes commit commit
| 1
|
46,678
| 7,275,022,638
|
IssuesEvent
|
2018-02-21 12:06:30
|
rucio/rucio
|
https://api.github.com/repos/rucio/rucio
|
closed
|
Broken readthedocs compilation with No such file or directory
|
Documentation
|
Motivation
----------
```
File "/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/envs/latest/local/lib/python2.7/site-packages/rucio-1.14.10-py2.7.egg/rucio/common/doc/argparse/ext.py", line 415, in run
f = open(os.path.abspath(self.options['filename']))
IOError: [Errno 2] No such file or directory: u'/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/checkouts/latest/doc/source/bin/rucio'
```
|
1.0
|
Broken readthedocs compilation with No such file or directory - Motivation
----------
```
File "/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/envs/latest/local/lib/python2.7/site-packages/rucio-1.14.10-py2.7.egg/rucio/common/doc/argparse/ext.py", line 415, in run
f = open(os.path.abspath(self.options['filename']))
IOError: [Errno 2] No such file or directory: u'/home/docs/checkouts/readthedocs.org/user_builds/rucio-blueprint/checkouts/latest/doc/source/bin/rucio'
```
|
non_process
|
broken readthedocs compilation with no such file or directory motivation file home docs checkouts readthedocs org user builds rucio blueprint envs latest local lib site packages rucio egg rucio common doc argparse ext py line in run f open os path abspath self options ioerror no such file or directory u home docs checkouts readthedocs org user builds rucio blueprint checkouts latest doc source bin rucio
| 0
|
324,214
| 9,886,276,587
|
IssuesEvent
|
2019-06-25 06:13:23
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.ebay.com - site is not usable
|
browser-firefox-mobile engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.ebay.com/i/303172994264?chn=ps&ul_ref=https%253A%252F%252Frover.ebay.com%252Frover%252F1%252F711-117182-37290-0%252F2%253Fmpre%253Dhttps%253A%252F%252Fwww.ebay.com%252Fi%252F303172994264%253Fchn%253Dps%2526var%253D602132596968%2526mkevt%253D1%2526mkrid%253D711-117182-37290-0%2526mkcid%253D2%2526itemid%253D602132596968_303172994264%2526targetid%253D503482154844%2526device%253Dm%2526adtype%253Dpla%2526googleloc%253D9029718%2526poi%253D%2526campaignid%253D1669934822%2526adgroupid%253D65058350059%2526rlsatarget%253Dpla-503482154844%2526abcId%253D1139296%2526merchantid%253D6296724%2526gclid%253DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%2526srcrot%253D711-117182-37290-0%2526rvr_id%253D2017044543838%2526rvr_ts%253D80b9d7e616b0aa48ddd4da87fff0994b
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: auto direct to ebay://link?...
**Steps to Reproduce**:
Even in desktop mode whole page forwarded to ebay://link?nav=item.view&id=303172994264&referrer=https%3A%2F%2Frover.ebay.com%2Frover%2F1%2F711-117182-37290-0%2F2%3Fvar%3D602132596968%26mkevt%3D1%26mkrid%3D711-117182-37290-0%26mkcid%3D2%26itemid%3D602132596968_303172994264%26targetid%3D503482154844%26device%3Dm%26adtype%3Dpla%26googleloc%3D9029718%26poi%3D%26campaignid%3D1669934822%26adgroupid%3D65058350059%26rlsatarget%3Dpla-503482154844%26abcId%3D1139296%26merchantid%3D6296724%26gclid%3DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%26srcrot%3D711-117182-37290-0%26rvr_id%3D2017044543838%26rvr_ts%3D80b9d7e616b0aa48ddd4da87fff0994b
I don't have the app installed, so it doesn't load.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190617221408</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>hasFastClick: true</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.ebay.com - site is not usable - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.ebay.com/i/303172994264?chn=ps&ul_ref=https%253A%252F%252Frover.ebay.com%252Frover%252F1%252F711-117182-37290-0%252F2%253Fmpre%253Dhttps%253A%252F%252Fwww.ebay.com%252Fi%252F303172994264%253Fchn%253Dps%2526var%253D602132596968%2526mkevt%253D1%2526mkrid%253D711-117182-37290-0%2526mkcid%253D2%2526itemid%253D602132596968_303172994264%2526targetid%253D503482154844%2526device%253Dm%2526adtype%253Dpla%2526googleloc%253D9029718%2526poi%253D%2526campaignid%253D1669934822%2526adgroupid%253D65058350059%2526rlsatarget%253Dpla-503482154844%2526abcId%253D1139296%2526merchantid%253D6296724%2526gclid%253DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%2526srcrot%253D711-117182-37290-0%2526rvr_id%253D2017044543838%2526rvr_ts%253D80b9d7e616b0aa48ddd4da87fff0994b
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: auto direct to ebay://link?...
**Steps to Reproduce**:
Even in desktop mode whole page forwarded to ebay://link?nav=item.view&id=303172994264&referrer=https%3A%2F%2Frover.ebay.com%2Frover%2F1%2F711-117182-37290-0%2F2%3Fvar%3D602132596968%26mkevt%3D1%26mkrid%3D711-117182-37290-0%26mkcid%3D2%26itemid%3D602132596968_303172994264%26targetid%3D503482154844%26device%3Dm%26adtype%3Dpla%26googleloc%3D9029718%26poi%3D%26campaignid%3D1669934822%26adgroupid%3D65058350059%26rlsatarget%3Dpla-503482154844%26abcId%3D1139296%26merchantid%3D6296724%26gclid%3DCj0KCQjwo7foBRD8ARIsAHTy2wnXcrl3yDjJIqzpViRqTX7arHPJGlEJMUJ_2olUYK7J6u3e5tjL7iUaAsVtEALw_wcB%26srcrot%3D711-117182-37290-0%26rvr_id%3D2017044543838%26rvr_ts%3D80b9d7e616b0aa48ddd4da87fff0994b
I don't have the app installed, so it doesn't load.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190617221408</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>hasFastClick: true</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description auto direct to ebay link steps to reproduce even in desktop mode whole page forwarded to ebay link nav item view id referrer https ebay com wcb id ts i don t have the app installed so it doesn t load browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false hasfastclick true gfx webrender enabled false gfx webrender all false channel beta from with ❤️
| 0
|
20,589
| 27,252,294,243
|
IssuesEvent
|
2023-02-22 09:03:27
|
haddocking/haddock3
|
https://api.github.com/repos/haddocking/haddock3
|
closed
|
add a postprocessing key to execute haddock3-analyse at the end of the workflow
|
analysis/postprocessing
|
For the workflow builder it would be nice to have `haddock3-analyse` run always at the end of the workflow. With a new option in the .cfg workflow (`postprocessing = true` ? ), haddock3 should run the postprocessing on a certain subset of folders (I would say only on the `caprieval` folders). In the workflow builder this option will be always active, while standard haddock command line users will have it set to false by default.
Discussed with @sverhoeven
|
1.0
|
add a postprocessing key to execute haddock3-analyse at the end of the workflow - For the workflow builder it would be nice to have `haddock3-analyse` run always at the end of the workflow. With a new option in the .cfg workflow (`postprocessing = true` ? ), haddock3 should run the postprocessing on a certain subset of folders (I would say only on the `caprieval` folders). In the workflow builder this option will be always active, while standard haddock command line users will have it set to false by default.
Discussed with @sverhoeven
|
process
|
add a postprocessing key to execute analyse at the end of the workflow for the workflow builder it would be nice to have analyse run always at the end of the workflow with a new option in the cfg workflow postprocessing true should run the postprocessing on a certain subset of folders i would say only on the caprieval folders in the workflow builder this option will be always active while standard haddock command line users will have it set to false by default discussed with sverhoeven
| 1
|
1,255
| 3,789,756,707
|
IssuesEvent
|
2016-03-21 19:01:25
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
opened
|
ProcessTests.TestExitTime failed on OSX
|
System.Diagnostics.Process
|
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/osx_debug_tst_prtest/2237/console
```
11:35:29 System.Diagnostics.Tests.ProcessTests.TestExitTime [FAIL]
11:35:29 Assert+WrapperXunitException : File path: d:\j\workspace\osx_debug_bld---2f4069d6\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 164
11:35:29 ---- Assert.True() Failure
11:35:29 Expected: True
11:35:29 Actual: False
11:35:29 Stack Trace:
11:35:29 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber)
11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
11:35:29 at System.Diagnostics.Tests.ProcessTests.TestExitTime()
11:35:29 ----- Inner Stack Trace -----
11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
```
|
1.0
|
ProcessTests.TestExitTime failed on OSX - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/osx_debug_tst_prtest/2237/console
```
11:35:29 System.Diagnostics.Tests.ProcessTests.TestExitTime [FAIL]
11:35:29 Assert+WrapperXunitException : File path: d:\j\workspace\osx_debug_bld---2f4069d6\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 164
11:35:29 ---- Assert.True() Failure
11:35:29 Expected: True
11:35:29 Actual: False
11:35:29 Stack Trace:
11:35:29 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber)
11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
11:35:29 at System.Diagnostics.Tests.ProcessTests.TestExitTime()
11:35:29 ----- Inner Stack Trace -----
11:35:29 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
```
|
process
|
processtests testexittime failed on osx system diagnostics tests processtests testexittime assert wrapperxunitexception file path d j workspace osx debug bld src system diagnostics process tests processtests cs line assert true failure expected true actual false stack trace at assert wrapexception exception e string callerfilepath callerlinenumber at assert true boolean condition string usermessage string path line at system diagnostics tests processtests testexittime inner stack trace at assert true boolean condition string usermessage string path line
| 1
|
1,297
| 3,837,908,645
|
IssuesEvent
|
2016-04-02 02:10:17
|
BayoAdejare/Leverage
|
https://api.github.com/repos/BayoAdejare/Leverage
|
opened
|
Create additional columns to cater testing and etc
|
process/administration
|
Create additional columns to cater testing and etc
|
1.0
|
Create additional columns to cater testing and etc - Create additional columns to cater testing and etc
|
process
|
create additional columns to cater testing and etc create additional columns to cater testing and etc
| 1
|
73,648
| 15,278,300,224
|
IssuesEvent
|
2021-02-23 01:09:35
|
emilwareus/thimble.mozilla.org
|
https://api.github.com/repos/emilwareus/thimble.mozilla.org
|
opened
|
CVE-2020-28499 (Medium) detected in merge-1.2.1.tgz
|
security vulnerability
|
## CVE-2020-28499 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary>
<p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p>
<p>Path to dependency file: thimble.mozilla.org/services/id.webmaker.org/package.json</p>
<p>Path to vulnerable library: thimble.mozilla.org/services/id.webmaker.org/node_modules/merge/package.json</p>
<p>
Dependency Hierarchy:
- autoless-0.1.7.tgz (Root Library)
- watch-0.14.0.tgz
- exec-sh-0.2.2.tgz
- :x: **merge-1.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/emilwareus/thimble.mozilla.org/commit/af3d91f99628f029ddcc04f2c30b6bf019be57d7">af3d91f99628f029ddcc04f2c30b6bf019be57d7</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-28499 (Medium) detected in merge-1.2.1.tgz - ## CVE-2020-28499 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary>
<p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p>
<p>Path to dependency file: thimble.mozilla.org/services/id.webmaker.org/package.json</p>
<p>Path to vulnerable library: thimble.mozilla.org/services/id.webmaker.org/node_modules/merge/package.json</p>
<p>
Dependency Hierarchy:
- autoless-0.1.7.tgz (Root Library)
- watch-0.14.0.tgz
- exec-sh-0.2.2.tgz
- :x: **merge-1.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/emilwareus/thimble.mozilla.org/commit/af3d91f99628f029ddcc04f2c30b6bf019be57d7">af3d91f99628f029ddcc04f2c30b6bf019be57d7</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in merge tgz cve medium severity vulnerability vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file thimble mozilla org services id webmaker org package json path to vulnerable library thimble mozilla org services id webmaker org node modules merge package json dependency hierarchy autoless tgz root library watch tgz exec sh tgz x merge tgz vulnerable library found in head commit a href vulnerability details all versions of package merge are vulnerable to prototype pollution via recursivemerge publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href step up your open source security game with whitesource
| 0
|
42,757
| 22,824,582,071
|
IssuesEvent
|
2022-07-12 07:24:59
|
Azure/azure-sdk-tools
|
https://api.github.com/repos/Azure/azure-sdk-tools
|
opened
|
[Perf] Move perf.yml into eng/common
|
pillar-performance
|
Most of `/eng/pipelines/templates/jobs/perf.yml` is duplicated across languages, and should be refactored into `/eng/common`.
|
True
|
[Perf] Move perf.yml into eng/common - Most of `/eng/pipelines/templates/jobs/perf.yml` is duplicated across languages, and should be refactored into `/eng/common`.
|
non_process
|
move perf yml into eng common most of eng pipelines templates jobs perf yml is duplicated across languages and should be refactored into eng common
| 0
|
18,229
| 24,296,427,727
|
IssuesEvent
|
2022-09-29 10:25:51
|
quark-engine/quark-engine
|
https://api.github.com/repos/quark-engine/quark-engine
|
closed
|
macOS Dependencies
|
test-required issue-processing-state-01
|
Dependencies error during the installation of the quark-engine on macOS Catalina - 10.15.7.
Error : -
pkg_resources.DistributionNotFound: The 'androguard==3.4.0a1' distribution was not found and is required by quark-engine
<img width="1016" alt="Screenshot 2020-11-02 at 6 18 56 PM" src="https://user-images.githubusercontent.com/18115054/97869748-fbcf3680-1d37-11eb-82e5-9132b3d03782.png">
|
1.0
|
macOS Dependencies - Dependencies error during the installation of the quark-engine on macOS Catalina - 10.15.7.
Error : -
pkg_resources.DistributionNotFound: The 'androguard==3.4.0a1' distribution was not found and is required by quark-engine
<img width="1016" alt="Screenshot 2020-11-02 at 6 18 56 PM" src="https://user-images.githubusercontent.com/18115054/97869748-fbcf3680-1d37-11eb-82e5-9132b3d03782.png">
|
process
|
macos dependencies dependencies error during the installation of the quark engine on macos catalina error pkg resources distributionnotfound the androguard distribution was not found and is required by quark engine img width alt screenshot at pm src
| 1
|
7,997
| 11,188,128,693
|
IssuesEvent
|
2020-01-02 02:55:43
|
52ABP/Documents
|
https://api.github.com/repos/52ABP/Documents
|
opened
|
ASP.NET Core 进程外(out-of-process)托管 | 52ABP官方技术文档与博客
|
ASP.NET Core 进程外(out-of-process)托管 | 52ABP官方技术文档与博 Gitalk
|
https://docs.52abp.com/mvc/7-Out-of-ProcessHosting.html
旨在打造新手小白从入门到实战的学习网站,内容涵盖:ASP.NET Core、Angular 、.NET Core、52ABP、等企业级解决方案等
|
1.0
|
ASP.NET Core 进程外(out-of-process)托管 | 52ABP官方技术文档与博客 - https://docs.52abp.com/mvc/7-Out-of-ProcessHosting.html
旨在打造新手小白从入门到实战的学习网站,内容涵盖:ASP.NET Core、Angular 、.NET Core、52ABP、等企业级解决方案等
|
process
|
asp net core 进程外 out of process 托管 旨在打造新手小白从入门到实战的学习网站,内容涵盖:asp net core、angular 、 net core、 、等企业级解决方案等
| 1
|
66,188
| 8,887,044,698
|
IssuesEvent
|
2019-01-15 03:31:04
|
NethermindEth/nethermind
|
https://api.github.com/repos/NethermindEth/nethermind
|
closed
|
Ropsten sync issues / discussion
|
documentation
|
Hi Tomasz,
im trying to run the Runner from within visual studio 2017. Im getting this error message:
_2018-11-06 09:41:48.7958|ERROR|1|Runner.Program|Failure|Config file does not exist configs\ropsten_windows_discovery.config.json_
Are the discovery files missing from the repo?
Thanks,
Anders.
|
1.0
|
Ropsten sync issues / discussion - Hi Tomasz,
im trying to run the Runner from within visual studio 2017. Im getting this error message:
_2018-11-06 09:41:48.7958|ERROR|1|Runner.Program|Failure|Config file does not exist configs\ropsten_windows_discovery.config.json_
Are the discovery files missing from the repo?
Thanks,
Anders.
|
non_process
|
ropsten sync issues discussion hi tomasz im trying to run the runner from within visual studio im getting this error message error runner program failure config file does not exist configs ropsten windows discovery config json are the discovery files missing from the repo thanks anders
| 0
|
288,298
| 24,897,082,037
|
IssuesEvent
|
2022-10-28 16:50:07
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts - Reporting APIs Usage from new jobs posted should handle preserve_layout pdf
|
Feature:Reporting failed-test Team:Global Experience
|
A test failed on a tracked branch
```
Error: expected 500 to equal 200
at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11)
at Assertion.be.Assertion.equal (node_modules/@kbn/expect/expect.js:227:8)
at Assertion.be (node_modules/@kbn/expect/expect.js:69:22)
at Object.waitForJobToFinish (x-pack/test/reporting_api_integration/services/usage.ts:55:31)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-f408bbae8ca11f13/elastic/kibana-on-merge/kibana/x-pack/test/reporting_api_integration/services/usage.ts:98:11
at async Promise.all (index 1)
at Object.expectAllJobsToFinishSuccessfully (x-pack/test/reporting_api_integration/services/usage.ts:96:7)
at Context.<anonymous> (x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs.ts:50:7)
at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
```
First failure: [CI Build - 8.3](https://buildkite.com/elastic/kibana-on-merge/builds/19791#01828ea3-05b6-4d80-87c9-4b633ecd8816)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts","test.name":"Reporting APIs Usage from new jobs posted should handle preserve_layout pdf","test.failCount":1}} -->
|
1.0
|
Failing test: X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts - Reporting APIs Usage from new jobs posted should handle preserve_layout pdf - A test failed on a tracked branch
```
Error: expected 500 to equal 200
at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11)
at Assertion.be.Assertion.equal (node_modules/@kbn/expect/expect.js:227:8)
at Assertion.be (node_modules/@kbn/expect/expect.js:69:22)
at Object.waitForJobToFinish (x-pack/test/reporting_api_integration/services/usage.ts:55:31)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-f408bbae8ca11f13/elastic/kibana-on-merge/kibana/x-pack/test/reporting_api_integration/services/usage.ts:98:11
at async Promise.all (index 1)
at Object.expectAllJobsToFinishSuccessfully (x-pack/test/reporting_api_integration/services/usage.ts:96:7)
at Context.<anonymous> (x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs.ts:50:7)
at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
```
First failure: [CI Build - 8.3](https://buildkite.com/elastic/kibana-on-merge/builds/19791#01828ea3-05b6-4d80-87c9-4b633ecd8816)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Reporting API Integration Tests.x-pack/test/reporting_api_integration/reporting_and_security/usage/new_jobs·ts","test.name":"Reporting APIs Usage from new jobs posted should handle preserve_layout pdf","test.failCount":1}} -->
|
non_process
|
failing test x pack reporting api integration tests x pack test reporting api integration reporting and security usage new jobs·ts reporting apis usage from new jobs posted should handle preserve layout pdf a test failed on a tracked branch error expected to equal at assertion assert node modules kbn expect expect js at assertion be assertion equal node modules kbn expect expect js at assertion be node modules kbn expect expect js at object waitforjobtofinish x pack test reporting api integration services usage ts at processticksandrejections node internal process task queues at var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack test reporting api integration services usage ts at async promise all index at object expectalljobstofinishsuccessfully x pack test reporting api integration services usage ts at context x pack test reporting api integration reporting and security usage new jobs ts at object apply node modules kbn test target node functional test runner lib mocha wrap function js first failure
| 0
|
587,065
| 17,603,669,756
|
IssuesEvent
|
2021-08-17 14:37:27
|
kirbydesign/designsystem
|
https://api.github.com/repos/kirbydesign/designsystem
|
closed
|
[Enhancement] Only keep 1 destructive button (Filled)
|
enhancement component:Button workaround possible priority 1 stale
|
**Is your enhancement request related to a problem? Please describe.**
In the use-cases we have right now we only have a filled destruction button (red background, black text)
We might also have an outline version that has a red outline with a black text, but thats it for now. The problem with a red text is that it is not accessible.
see also: https://zpl.io/VDKnZAl
|
1.0
|
[Enhancement] Only keep 1 destructive button (Filled) - **Is your enhancement request related to a problem? Please describe.**
In the use-cases we have right now we only have a filled destruction button (red background, black text)
We might also have an outline version that has a red outline with a black text, but thats it for now. The problem with a red text is that it is not accessible.
see also: https://zpl.io/VDKnZAl
|
non_process
|
only keep destructive button filled is your enhancement request related to a problem please describe in the use cases we have right now we only have a filled destruction button red background black text we might also have an outline version that has a red outline with a black text but thats it for now the problem with a red text is that it is not accessible see also
| 0
|
560
| 3,021,629,256
|
IssuesEvent
|
2015-07-31 15:45:43
|
18F/CMS.gov-developer
|
https://api.github.com/repos/18F/CMS.gov-developer
|
opened
|
Plan training dates
|
process
|
* [ ] CMS to consider what kind of trainings and with you.
* [ ] CMS to pick tentative dates and times.
|
1.0
|
Plan training dates - * [ ] CMS to consider what kind of trainings and with you.
* [ ] CMS to pick tentative dates and times.
|
process
|
plan training dates cms to consider what kind of trainings and with you cms to pick tentative dates and times
| 1
|
12,220
| 14,743,131,475
|
IssuesEvent
|
2021-01-07 13:26:54
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Towne - Invoice changes?
|
anc-process anp-1.5 ant-support
|
In GitLab by @kdjstudios on Jul 16, 2019, 09:42
**Submitted by:** Deb Crown <dcrown@towneanswering.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8806788
**Server:** External
**Client/Site:** Towne
**Account:** NA
**Issue:**
I guess there was some work done behind the scenes that created a situation that – instead of printing Late Fee as a line item on the invoice – it now simply prints 1.50%
Also, SA generated, printed and/or emailed invoices that have a zero balance. I don’t think it did that before, did it?
|
1.0
|
Towne - Invoice changes? - In GitLab by @kdjstudios on Jul 16, 2019, 09:42
**Submitted by:** Deb Crown <dcrown@towneanswering.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8806788
**Server:** External
**Client/Site:** Towne
**Account:** NA
**Issue:**
I guess there was some work done behind the scenes that created a situation that – instead of printing Late Fee as a line item on the invoice – it now simply prints 1.50%
Also, SA generated, printed and/or emailed invoices that have a zero balance. I don’t think it did that before, did it?
|
process
|
towne invoice changes in gitlab by kdjstudios on jul submitted by deb crown helpdesk server external client site towne account na issue i guess there was some work done behind the scenes that created a situation that – instead of printing late fee as a line item on the invoice – it now simply prints also sa generated printed and or emailed invoices that have a zero balance i don’t think it did that before did it
| 1
|
5,410
| 8,245,815,590
|
IssuesEvent
|
2018-09-11 10:52:57
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[needs-docs][processing] return empty layer for save
selected features algorithm if no features are selected
|
Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/0cdabb5e60bd71941b559bf4f74e75e423b56536 by nirvn
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[needs-docs][processing] return empty layer for save
selected features algorithm if no features are selected - Original commit: https://github.com/qgis/QGIS/commit/0cdabb5e60bd71941b559bf4f74e75e423b56536 by nirvn
Unfortunately this naughty coder did not write a description... :-(
|
process
|
return empty layer for save selected features algorithm if no features are selected original commit by nirvn unfortunately this naughty coder did not write a description
| 1
|
9,569
| 12,519,746,233
|
IssuesEvent
|
2020-06-03 14:52:31
|
GoogleCloudPlatform/dotnet-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
|
opened
|
[Monitoring] Skip some of the tests because they are doing cuncurrent modifications that fail.
|
api: monitoring priority: p1 type: process
|
Even thouhg in #1067 each test project was configured to not run tests in parallel, the different projects can still be being run in parallel and also we are running tests in Windows and Linux, which run roughly at the same time on presubmit. Several of the test projects are using the same fixture, and there is where we get "Too many concurrent edits to the project configuration". For fixing this we'll need to:
- Merge all test projects into one and use the fixture as a collection fixture.
- Add a retry on the fixture when creating the resources in case Windows and Linux are running at the same time.
Tests will be disabled in #1067 until these fixes can be made.
|
1.0
|
[Monitoring] Skip some of the tests because they are doing cuncurrent modifications that fail. - Even thouhg in #1067 each test project was configured to not run tests in parallel, the different projects can still be being run in parallel and also we are running tests in Windows and Linux, which run roughly at the same time on presubmit. Several of the test projects are using the same fixture, and there is where we get "Too many concurrent edits to the project configuration". For fixing this we'll need to:
- Merge all test projects into one and use the fixture as a collection fixture.
- Add a retry on the fixture when creating the resources in case Windows and Linux are running at the same time.
Tests will be disabled in #1067 until these fixes can be made.
|
process
|
skip some of the tests because they are doing cuncurrent modifications that fail even thouhg in each test project was configured to not run tests in parallel the different projects can still be being run in parallel and also we are running tests in windows and linux which run roughly at the same time on presubmit several of the test projects are using the same fixture and there is where we get too many concurrent edits to the project configuration for fixing this we ll need to merge all test projects into one and use the fixture as a collection fixture add a retry on the fixture when creating the resources in case windows and linux are running at the same time tests will be disabled in until these fixes can be made
| 1
|
490,066
| 14,115,102,726
|
IssuesEvent
|
2020-11-07 19:09:54
|
HHS81/c182s
|
https://api.github.com/repos/HHS81/c182s
|
closed
|
use `/fdm/jsbsim/propulsion/engine/AFR` for exhaust smoke in rich operation
|
effects low priority refactoring wontfix
|
`/fdm/jsbsim/propulsion/engine/AFR` offers the Air-to-Fuel ratio in the current FGFS GIT version. This could be used to generate smoke from the exhaust when mixture is very rich and may also replace the naive start-smoke implementation from #297 (which occurs also because of very rich starter mixture).
|
1.0
|
use `/fdm/jsbsim/propulsion/engine/AFR` for exhaust smoke in rich operation - `/fdm/jsbsim/propulsion/engine/AFR` offers the Air-to-Fuel ratio in the current FGFS GIT version. This could be used to generate smoke from the exhaust when mixture is very rich and may also replace the naive start-smoke implementation from #297 (which occurs also because of very rich starter mixture).
|
non_process
|
use fdm jsbsim propulsion engine afr for exhaust smoke in rich operation fdm jsbsim propulsion engine afr offers the air to fuel ratio in the current fgfs git version this could be used to generate smoke from the exhaust when mixture is very rich and may also replace the naive start smoke implementation from which occurs also because of very rich starter mixture
| 0
|
284,091
| 8,735,673,193
|
IssuesEvent
|
2018-12-11 17:22:02
|
aowen87/TicketTester
|
https://api.github.com/repos/aowen87/TicketTester
|
closed
|
Gui crashes attempting to open on remote after using 'save session' dialog
|
bug likelihood medium priority reviewed severity high
|
Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2385
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Gui crashes attempting to open on remote after using 'save session' dialog
Assigned to:
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 09/22/2015
Due date:
% Done: 0
Estimated time:
Created: 09/22/2015 12:50 pm
Updated: 10/06/2015 08:40 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: trunk
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash.
Comments:
I tried to duplicate this problem on a Linux system, but could not duplicate the crash.But I am not sure if I really duplicated the problem because I have questions about the steps.Also Brad Whitlock made some changes to the session save code and this could have fixed the issue, but here are the steps I tried to duplicate the problem.1) Open a data file on the local machine. I opened the noise data set.2) Created a Mesh plot on the mesh variable.3) Open the "Save session" dialog and then clicked cancel.4) Here is were I am not sure what "Open" means, which dialog "Save session", "Restore session", or "Open file". I am also not sure what "select a remote host" means, does that mean clicked on the host drop down or field itself? So I tried the following: 1) Open the "Save session" dialog. Then click on the host field text box and selected the drop down. 2) Open the "Restore session" dialog. Then click on the host field text box and selected the drop down. 3) Open the "Open file" dialog. Then click on the host field text box and selected the drop down.None of these steps caused a crash. So did I miss something in duplicating the problem?Thanks,David 1) Open a file and draw a plot2) Open 'Save session' dialog, cancel it3) Click 'Open' to get the file open dialog, select the host dropdown and choose a remote hostGui will crash.gdb shows a segv from QvisFileWindowBase::ProgressCallback in the call toThis->setEnabled(false) Short way to duplicate the problem:1) Click the "Open" data button on GUI and click "Cancel" button.2) Click File->"Save session" to open dialog and click "Cancel" button.3) Click the "Open" data button on GUI and select different host in the host drop down.Crash inside Qt setEnabled_helper called by QvisFileWindowBase::ProgressCallback.So I think there must be something shared between the Open Data Window and the Session save dialog and when the Session Save dialog is deleted it must close or delete something they both share. Resolved by David Camp with SVN #27306, which was ported to 2.10RC with SVN #27402.
|
1.0
|
Gui crashes attempting to open on remote after using 'save session' dialog - Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2385
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Gui crashes attempting to open on remote after using 'save session' dialog
Assigned to:
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 09/22/2015
Due date:
% Done: 0
Estimated time:
Created: 09/22/2015 12:50 pm
Updated: 10/06/2015 08:40 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: trunk
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Open a file and draw a plot on the local machine. Open 'Save session' dialog, but don't do anything, just cancel. Open, select a remote host, the gui will crash.
Comments:
I tried to duplicate this problem on a Linux system, but could not duplicate the crash.But I am not sure if I really duplicated the problem because I have questions about the steps.Also Brad Whitlock made some changes to the session save code and this could have fixed the issue, but here are the steps I tried to duplicate the problem.1) Open a data file on the local machine. I opened the noise data set.2) Created a Mesh plot on the mesh variable.3) Open the "Save session" dialog and then clicked cancel.4) Here is were I am not sure what "Open" means, which dialog "Save session", "Restore session", or "Open file". I am also not sure what "select a remote host" means, does that mean clicked on the host drop down or field itself? So I tried the following: 1) Open the "Save session" dialog. Then click on the host field text box and selected the drop down. 2) Open the "Restore session" dialog. Then click on the host field text box and selected the drop down. 3) Open the "Open file" dialog. Then click on the host field text box and selected the drop down.None of these steps caused a crash. So did I miss something in duplicating the problem?Thanks,David 1) Open a file and draw a plot2) Open 'Save session' dialog, cancel it3) Click 'Open' to get the file open dialog, select the host dropdown and choose a remote hostGui will crash.gdb shows a segv from QvisFileWindowBase::ProgressCallback in the call toThis->setEnabled(false) Short way to duplicate the problem:1) Click the "Open" data button on GUI and click "Cancel" button.2) Click File->"Save session" to open dialog and click "Cancel" button.3) Click the "Open" data button on GUI and select different host in the host drop down.Crash inside Qt setEnabled_helper called by QvisFileWindowBase::ProgressCallback.So I think there must be something shared between the Open Data Window and the Session save dialog and when the Session Save dialog is deleted it must close or delete something they both share. Resolved by David Camp with SVN #27306, which was ported to 2.10RC with SVN #27402.
|
non_process
|
gui crashes attempting to open on remote after using save session dialog open a file and draw a plot on the local machine open save session dialog but don t do anything just cancel open select a remote host the gui will crash redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject gui crashes attempting to open on remote after using save session dialog assigned to category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version trunk impact expected use os all support group any description open a file and draw a plot on the local machine open save session dialog but don t do anything just cancel open select a remote host the gui will crash comments i tried to duplicate this problem on a linux system but could not duplicate the crash but i am not sure if i really duplicated the problem because i have questions about the steps also brad whitlock made some changes to the session save code and this could have fixed the issue but here are the steps i tried to duplicate the problem open a data file on the local machine i opened the noise data set created a mesh plot on the mesh variable open the save session dialog and then clicked cancel here is were i am not sure what open means which dialog save session restore session or open file i am also not sure what select a remote host means does that mean clicked on the host drop down or field itself so i tried the following open the save session dialog then click on the host field text box and selected the drop down open the restore session dialog then click on the host field text box and selected the drop down open the open file dialog then click on the host field text box and selected the drop down none of these steps caused a crash so did i miss something in duplicating the problem thanks david open a file and draw a open save session dialog cancel click open to get the file open dialog select the host dropdown and choose a remote hostgui will crash gdb shows a segv from qvisfilewindowbase progresscallback in the call tothis setenabled false short way to duplicate the problem click the open data button on gui and click cancel button click file save session to open dialog and click cancel button click the open data button on gui and select different host in the host drop down crash inside qt setenabled helper called by qvisfilewindowbase progresscallback so i think there must be something shared between the open data window and the session save dialog and when the session save dialog is deleted it must close or delete something they both share resolved by david camp with svn which was ported to with svn
| 0
|
224,674
| 24,783,421,931
|
IssuesEvent
|
2022-10-24 07:50:21
|
sast-automation-dev/openidm-community-edition-43
|
https://api.github.com/repos/sast-automation-dev/openidm-community-edition-43
|
opened
|
openidm-config-2.1.2.jar: 1 vulnerabilities (highest severity is: 6.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openidm-config-2.1.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /openidm-shell/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (openidm-config version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [WS-2019-0379](https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | commons-codec-1.7.jar | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0379</summary>
### Vulnerable Library - <b>commons-codec-1.7.jar</b></p>
<p>The codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Library home page: <a href="http://commons.apache.org/codec/">http://commons.apache.org/codec/</a></p>
<p>Path to dependency file: /openidm-repo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p>
<p>
Dependency Hierarchy:
- openidm-config-2.1.2.jar (Root Library)
- openidm-repo-2.1.2.jar
- openidm-util-2.1.2.jar
- json-crypto-core-1.1.0.jar
- :x: **commons-codec-1.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: May 20, 2019 3:39:18 PM
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: May 20, 2019 3:39:18 PM</p>
<p>Fix Resolution: commons-codec:commons-codec:1.13</p>
</p>
<p></p>
</details>
|
True
|
openidm-config-2.1.2.jar: 1 vulnerabilities (highest severity is: 6.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openidm-config-2.1.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /openidm-shell/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (openidm-config version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [WS-2019-0379](https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | commons-codec-1.7.jar | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0379</summary>
### Vulnerable Library - <b>commons-codec-1.7.jar</b></p>
<p>The codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Library home page: <a href="http://commons.apache.org/codec/">http://commons.apache.org/codec/</a></p>
<p>Path to dependency file: /openidm-repo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar</p>
<p>
Dependency Hierarchy:
- openidm-config-2.1.2.jar (Root Library)
- openidm-repo-2.1.2.jar
- openidm-util-2.1.2.jar
- json-crypto-core-1.1.0.jar
- :x: **commons-codec-1.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openidm-community-edition-43/commit/0aad6d987ba225eeadc591c7c188b6deef985e1b">0aad6d987ba225eeadc591c7c188b6deef985e1b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: May 20, 2019 3:39:18 PM
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: May 20, 2019 3:39:18 PM</p>
<p>Fix Resolution: commons-codec:commons-codec:1.13</p>
</p>
<p></p>
</details>
|
non_process
|
openidm config jar vulnerabilities highest severity is vulnerable library openidm config jar path to dependency file openidm shell pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in openidm config version remediation available medium commons codec jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details ws vulnerable library commons codec jar the codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities library home page a href path to dependency file openidm repo pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy openidm config jar root library openidm repo jar openidm util jar json crypto core jar x commons codec jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date may pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date may pm fix resolution commons codec commons codec
| 0
|
22,494
| 31,468,484,843
|
IssuesEvent
|
2023-08-30 05:19:51
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
When save questions, the placeholder name is not consistent
|
Type:Bug Priority:P3 .Backend Querying/GUI .metabase-lib .Team/QueryProcessor :hammer_and_wrench: .Wanted: MLv2
|
**Describe the bug**
When saving a question, we automatically fill out a name for it based on the source table name.
But when a question has a join, the name does not contain the joined table.
**To Reproduce**
1. New Question
2. Sample Dataset -> Orders
3. Join with the Products table
4. Click save

5. Notice the name in the form (1st arrow) is "Orders", But the name of the question (2nd arrow) is "Orders + Products" which I think is a neat auto format.
**Expected behavior**
The name should be "Orders + Products"
**Severity**
Minor
**Additional context**
Tested with current master(1ca62d1b43cc97dc9af4886fa0abee8aa025a47c)
|
1.0
|
When save questions, the placeholder name is not consistent - **Describe the bug**
When saving a question, we automatically fill out a name for it based on the source table name.
But when a question has a join, the name does not contain the joined table.
**To Reproduce**
1. New Question
2. Sample Dataset -> Orders
3. Join with the Products table
4. Click save

5. Notice the name in the form (1st arrow) is "Orders", But the name of the question (2nd arrow) is "Orders + Products" which I think is a neat auto format.
**Expected behavior**
The name should be "Orders + Products"
**Severity**
Minor
**Additional context**
Tested with current master(1ca62d1b43cc97dc9af4886fa0abee8aa025a47c)
|
process
|
when save questions the placeholder name is not consistent describe the bug when saving a question we automatically fill out a name for it based on the source table name but when a question has a join the name does not contain the joined table to reproduce new question sample dataset orders join with the products table click save notice the name in the form arrow is orders but the name of the question arrow is orders products which i think is a neat auto format expected behavior the name should be orders products severity minor additional context tested with current master
| 1
|
83,381
| 3,634,281,793
|
IssuesEvent
|
2016-02-11 17:24:59
|
PCMDI/pcmdi_metrics
|
https://api.github.com/repos/PCMDI/pcmdi_metrics
|
reopened
|
output clim not masked for ocean case...
|
bug TOPPRIORITY
|
In the parameter file, we can select variables for which we want metrics on 'global', 'ocean' and 'land' case (e.g., below). Global and ocean cases are ok, but the output.nc clims for the 'ocean' case are not masked over land.
## REGIONS ON WHICH WE WANT TO RUN METRICS (var specific)
regions = {"tas" : [None,"land","ocean"],
"uas" : [None,"land","ocean"],
"vas" : [None,"land","ocean"],
"pr" : [None,"land","ocean"],
"psl": [None,"land","ocean",],
"huss": [None,"land","ocean"],
"prw": [None,"land","ocean"],
|
1.0
|
output clim not masked for ocean case... - In the parameter file, we can select variables for which we want metrics on 'global', 'ocean' and 'land' case (e.g., below). Global and ocean cases are ok, but the output.nc clims for the 'ocean' case are not masked over land.
## REGIONS ON WHICH WE WANT TO RUN METRICS (var specific)
regions = {"tas" : [None,"land","ocean"],
"uas" : [None,"land","ocean"],
"vas" : [None,"land","ocean"],
"pr" : [None,"land","ocean"],
"psl": [None,"land","ocean",],
"huss": [None,"land","ocean"],
"prw": [None,"land","ocean"],
|
non_process
|
output clim not masked for ocean case in the parameter file we can select variables for which we want metrics on global ocean and land case e g below global and ocean cases are ok but the output nc clims for the ocean case are not masked over land regions on which we want to run metrics var specific regions tas uas vas pr psl huss prw
| 0
|
175,817
| 27,979,244,304
|
IssuesEvent
|
2023-03-26 00:16:28
|
kocheck/Lazy-Export
|
https://api.github.com/repos/kocheck/Lazy-Export
|
closed
|
New Branding
|
Needs Design no-issue-activity
|
## Overview
Let's be honest, we need a logo and Artwork created. I have been putting this off and I really should dig deep here for something fun.
|
1.0
|
New Branding - ## Overview
Let's be honest, we need a logo and Artwork created. I have been putting this off and I really should dig deep here for something fun.
|
non_process
|
new branding overview let s be honest we need a logo and artwork created i have been putting this off and i really should dig deep here for something fun
| 0
|
7,405
| 10,523,941,063
|
IssuesEvent
|
2019-09-30 12:15:14
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
closed
|
Lift broken without init flow
|
process/candidate
|
```sh
$ ./node_modules/.bin/prisma2 introspect --pg-host localhost --pg-password "" --pg-user m --pg-schema public --pg-db prisma-blog
$ mv datamodel* schema.prisma
$ ./node_modules/.bin/prisma2 lift up
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.:
Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5
stack backtrace:
0: std::panicking::default_hook::{{closure}}
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.:
Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5
stack backtrace:
0: std::panicking::default_hook::{{closure}}
1: std::panicking::default_hook
2: migration_engine::main::{{closure}}
3: std::panicking::rust_panic_with_hook
4: std::panicking::continue_panic_fmt
5: rust_begin_unwind
6: core::panicking::panic_fmt
7: core::result::unwrap_failed
8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::deserialize_databas
e_migration
9: <migration_engine::commands::list_migrations::ListMigrationStepsCommand as migration_engine::commands::command::
MigrationCommand>::execute
10: <migration_engine::api::MigrationApi<C,D> as migration_engine::api::GenericApi>::list_migrations
11: migration_engine::api::rpc::RpcApi::create_sync_handler
12: tokio_executor::enter::exit
13: tokio_threadpool::blocking::blocking
14: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll
15: futures::future::chain::Chain<A,B,C>::poll
16: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll
17: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll
18: futures::future::chain::Chain<A,B,C>::poll
19: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll
20: <futures::future::map::Map<A,F> as futures::future::Future>::poll
21: <futures::future::either::Either<A,B> as futures::future::Future>::poll
22: <futures::future::map::Map<A,F> as futures::future::Future>::poll
23: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll
24: <futures::stream::and_then::AndThen<S,F,U> as futures::stream::Stream>::poll
25: <futures::stream::forward::Forward<T,U> as futures::future::Future>::poll
26: <futures::future::map::Map<A,F> as futures::future::Future>::poll
27: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll
28: futures::task_impl::std::set
29: std::panicking::try::do_call
30: __rust_maybe_catch_panic
31: tokio_threadpool::task::Task::run
32: tokio_threadpool::worker::Worker::run_task
33: tokio_threadpool::worker::Worker::run
34: std::thread::local::LocalKey<T>::with
35: std::thread::local::LocalKey<T>::with
36: tokio_reactor::with_default
37: tokio::runtime::threadpool::builder::Builder::build::{{closure}}
38: std::thread::local::LocalKey<T>::with
39: std::thread::local::LocalKey<T>::with
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
Please help us improve Prisma 2 by submitting an error report.
Error reports never contain personal or other sensitive information.
Learn more: https://pris.ly/d/telemetry
```
|
1.0
|
Lift broken without init flow - ```sh
$ ./node_modules/.bin/prisma2 introspect --pg-host localhost --pg-password "" --pg-user m --pg-schema public --pg-db prisma-blog
$ mv datamodel* schema.prisma
$ ./node_modules/.bin/prisma2 lift up
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.:
Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5
stack backtrace:
0: std::panicking::default_hook::{{closure}}
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'tokio-runtime-worker-0' panicked at 'Deserializing the database migration failed.:
Error("missing field `original_steps`", line: 0, column: 0)', src/libcore/result.rs:999:5
stack backtrace:
0: std::panicking::default_hook::{{closure}}
1: std::panicking::default_hook
2: migration_engine::main::{{closure}}
3: std::panicking::rust_panic_with_hook
4: std::panicking::continue_panic_fmt
5: rust_begin_unwind
6: core::panicking::panic_fmt
7: core::result::unwrap_failed
8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::deserialize_databas
e_migration
9: <migration_engine::commands::list_migrations::ListMigrationStepsCommand as migration_engine::commands::command::
MigrationCommand>::execute
10: <migration_engine::api::MigrationApi<C,D> as migration_engine::api::GenericApi>::list_migrations
11: migration_engine::api::rpc::RpcApi::create_sync_handler
12: tokio_executor::enter::exit
13: tokio_threadpool::blocking::blocking
14: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll
15: futures::future::chain::Chain<A,B,C>::poll
16: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll
17: <futures::future::lazy::Lazy<F,R> as futures::future::Future>::poll
18: futures::future::chain::Chain<A,B,C>::poll
19: <futures::future::then::Then<A,B,F> as futures::future::Future>::poll
20: <futures::future::map::Map<A,F> as futures::future::Future>::poll
21: <futures::future::either::Either<A,B> as futures::future::Future>::poll
22: <futures::future::map::Map<A,F> as futures::future::Future>::poll
23: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll
24: <futures::stream::and_then::AndThen<S,F,U> as futures::stream::Stream>::poll
25: <futures::stream::forward::Forward<T,U> as futures::future::Future>::poll
26: <futures::future::map::Map<A,F> as futures::future::Future>::poll
27: <futures::future::map_err::MapErr<A,F> as futures::future::Future>::poll
28: futures::task_impl::std::set
29: std::panicking::try::do_call
30: __rust_maybe_catch_panic
31: tokio_threadpool::task::Task::run
32: tokio_threadpool::worker::Worker::run_task
33: tokio_threadpool::worker::Worker::run
34: std::thread::local::LocalKey<T>::with
35: std::thread::local::LocalKey<T>::with
36: tokio_reactor::with_default
37: tokio::runtime::threadpool::builder::Builder::build::{{closure}}
38: std::thread::local::LocalKey<T>::with
39: std::thread::local::LocalKey<T>::with
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
Please help us improve Prisma 2 by submitting an error report.
Error reports never contain personal or other sensitive information.
Learn more: https://pris.ly/d/telemetry
```
|
process
|
lift broken without init flow sh node modules bin introspect pg host localhost pg password pg user m pg schema public pg db prisma blog mv datamodel schema prisma node modules bin lift up error oops an unexpected error occured error in migration engine thread tokio runtime worker panicked at deserializing the database migration failed error missing field original steps line column src libcore result rs stack backtrace std panicking default hook closure error oops an unexpected error occured error in migration engine thread tokio runtime worker panicked at deserializing the database migration failed error missing field original steps line column src libcore result rs stack backtrace std panicking default hook closure std panicking default hook migration engine main closure std panicking rust panic with hook std panicking continue panic fmt rust begin unwind core panicking panic fmt core result unwrap failed deserialize databas e migration migration engine commands list migrations listmigrationstepscommand as migration engine commands command migrationcommand execute as migration engine api genericapi list migrations migration engine api rpc rpcapi create sync handler tokio executor enter exit tokio threadpool blocking blocking as futures future future poll futures future chain chain poll as futures future future poll as futures future future poll futures future chain chain poll as futures future future poll as futures future future poll as futures future future poll as futures future future poll as futures future future poll as futures stream stream poll as futures future future poll as futures future future poll as futures future future poll futures task impl std set std panicking try do call rust maybe catch panic tokio threadpool task task run tokio threadpool worker worker run task tokio threadpool worker worker run std thread local localkey with std thread local localkey with tokio reactor with default tokio runtime threadpool builder builder build closure std thread local localkey with std thread local localkey with note some details are omitted run with rust backtrace full for a verbose backtrace please help us improve prisma by submitting an error report error reports never contain personal or other sensitive information learn more
| 1
|
361,035
| 10,702,851,520
|
IssuesEvent
|
2019-10-24 08:23:38
|
celo-org/celo-monorepo
|
https://api.github.com/repos/celo-org/celo-monorepo
|
opened
|
Analytics should be reported from the iOS app
|
Priority: P0 applications ios wallet
|
### Expected Behavior
Analytics are reported from the iOS app.
### Current Behavior
Analytics are not reported from the iOS app.
|
1.0
|
Analytics should be reported from the iOS app - ### Expected Behavior
Analytics are reported from the iOS app.
### Current Behavior
Analytics are not reported from the iOS app.
|
non_process
|
analytics should be reported from the ios app expected behavior analytics are reported from the ios app current behavior analytics are not reported from the ios app
| 0
|
10,223
| 13,093,700,514
|
IssuesEvent
|
2020-08-03 10:54:51
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
improvement of spatial selection
|
Feature Request Feedback Processing
|
Author Name: **jannekevandijk -** (jannekevandijk -)
Original Redmine Issue: [3703](https://issues.qgis.org/issues/3703)
Redmine category:c++_plugins/spatial_query
---
basically something similar to [[ArcGis]] 'select by location', where you
are able to:
1. select features in multiple layers
2. select using a buffer without having to create buffers first (eg select features within a distance of selected features of another layer)
(or should this be two seperate tickets?)
|
1.0
|
improvement of spatial selection - Author Name: **jannekevandijk -** (jannekevandijk -)
Original Redmine Issue: [3703](https://issues.qgis.org/issues/3703)
Redmine category:c++_plugins/spatial_query
---
basically something similar to [[ArcGis]] 'select by location', where you
are able to:
1. select features in multiple layers
2. select using a buffer without having to create buffers first (eg select features within a distance of selected features of another layer)
(or should this be two seperate tickets?)
|
process
|
improvement of spatial selection author name jannekevandijk jannekevandijk original redmine issue redmine category c plugins spatial query basically something similar to select by location where you are able to select features in multiple layers select using a buffer without having to create buffers first eg select features within a distance of selected features of another layer or should this be two seperate tickets
| 1
|
424,692
| 12,322,166,329
|
IssuesEvent
|
2020-05-13 09:53:27
|
ooni/probe-engine
|
https://api.github.com/repos/ooni/probe-engine
|
opened
|
cleanup: stop downloading CA as a resource
|
effort/XS priority/low
|
We cannot do that now, because MK code still needs it. We will be able to do so when we have removed all MK code. At that point, also the CABundlePath of the session can be removed.
This follows-up from https://github.com/ooni/probe-engine/issues/335
|
1.0
|
cleanup: stop downloading CA as a resource - We cannot do that now, because MK code still needs it. We will be able to do so when we have removed all MK code. At that point, also the CABundlePath of the session can be removed.
This follows-up from https://github.com/ooni/probe-engine/issues/335
|
non_process
|
cleanup stop downloading ca as a resource we cannot do that now because mk code still needs it we will be able to do so when we have removed all mk code at that point also the cabundlepath of the session can be removed this follows up from
| 0
|
15,562
| 19,703,503,948
|
IssuesEvent
|
2022-01-12 19:08:02
|
googleapis/java-security-private-ca
|
https://api.github.com/repos/googleapis/java-security-private-ca
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'security-private-ca' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'security-private-ca' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname security private ca invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
243,667
| 26,287,388,377
|
IssuesEvent
|
2023-01-08 01:04:09
|
Yoavmartin/spring-petclinic
|
https://api.github.com/repos/Yoavmartin/spring-petclinic
|
opened
|
CVE-2019-8331 (Medium) detected in bootstrap-3.3.6.jar
|
security vulnerability
|
## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.jar</b></p></summary>
<p>WebJar for Bootstrap</p>
<p>Library home page: <a href="http://webjars.org">http://webjars.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /canner/.m2/repository/org/webjars/bootstrap/3.3.6/bootstrap-3.3.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: 3.4.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
True
|
CVE-2019-8331 (Medium) detected in bootstrap-3.3.6.jar - ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.jar</b></p></summary>
<p>WebJar for Bootstrap</p>
<p>Library home page: <a href="http://webjars.org">http://webjars.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /canner/.m2/repository/org/webjars/bootstrap/3.3.6/bootstrap-3.3.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: 3.4.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
non_process
|
cve medium detected in bootstrap jar cve medium severity vulnerability vulnerable library bootstrap jar webjar for bootstrap library home page a href path to dependency file pom xml path to vulnerable library canner repository org webjars bootstrap bootstrap jar dependency hierarchy x bootstrap jar vulnerable library found in base branch main vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue
| 0
|
646,401
| 21,046,878,003
|
IssuesEvent
|
2022-03-31 16:50:14
|
Zydrax/Baldazzar-BUG-Traker
|
https://api.github.com/repos/Zydrax/Baldazzar-BUG-Traker
|
closed
|
Quest - Preemptive Methods (26873)
|
Medium Priority
|
Current behaviour: Serpentbloom has no loot
Expected behaviour: Serpentbloom should drop 1 Serpentbloom (26873)
Steps to reproduce the problem: Loot
Include proofs for this behaviour
https://www.wowhead.com/quest=26873/preemptive-methods
|
1.0
|
Quest - Preemptive Methods (26873) - Current behaviour: Serpentbloom has no loot
Expected behaviour: Serpentbloom should drop 1 Serpentbloom (26873)
Steps to reproduce the problem: Loot
Include proofs for this behaviour
https://www.wowhead.com/quest=26873/preemptive-methods
|
non_process
|
quest preemptive methods current behaviour serpentbloom has no loot expected behaviour serpentbloom should drop serpentbloom steps to reproduce the problem loot include proofs for this behaviour
| 0
|
4,658
| 7,496,686,023
|
IssuesEvent
|
2018-04-08 12:11:35
|
ODiogoSilva/assemblerflow
|
https://api.github.com/repos/ODiogoSilva/assemblerflow
|
closed
|
Add assembly process template for skesa
|
enhancement process
|
This template is akin to spade's `process_spades` template that is meant to parse and assembly by contig size and average kmer coverage.
|
1.0
|
Add assembly process template for skesa - This template is akin to spade's `process_spades` template that is meant to parse and assembly by contig size and average kmer coverage.
|
process
|
add assembly process template for skesa this template is akin to spade s process spades template that is meant to parse and assembly by contig size and average kmer coverage
| 1
|
64,044
| 12,413,786,736
|
IssuesEvent
|
2020-05-22 13:24:09
|
eclipse/che
|
https://api.github.com/repos/eclipse/che
|
closed
|
Git clone using private key with password is not working
|
area/che-theia kind/bug severity/P1 status/code-review
|
### Describe the bug
When trying to clone git repository using `git@gitlab.com/something/something` using ssh key, which is protected by password, Che never asks me for a password and the clone fails with this error:
```
git clone git@gitlab.something/something/web-app.git /projects/web-app
Cloning into '/projects/web-app'...
git@gitlab.com: Permission denied (publickey).
fatal: Could not read from remote repository.
Please make sure you have the correct access rights
and the repository exists.
```
### Che version
che.openshift.io
- [x] latest
- [ ] nightly
- [ ] other: please specify
### Steps to reproduce
* Generate ssh keypair *with a password* (`ssh-keygen`)
* Upload public part of that key to gitlab/github
* Start any workspace
* upload private part of the key to the workspace
* try to clone some repo using this format -> `git@gitlab.something/something/web-app.git`
### Expected behavior
Repo is cloned
|
1.0
|
Git clone using private key with password is not working - ### Describe the bug
When trying to clone git repository using `git@gitlab.com/something/something` using ssh key, which is protected by password, Che never asks me for a password and the clone fails with this error:
```
git clone git@gitlab.something/something/web-app.git /projects/web-app
Cloning into '/projects/web-app'...
git@gitlab.com: Permission denied (publickey).
fatal: Could not read from remote repository.
Please make sure you have the correct access rights
and the repository exists.
```
### Che version
che.openshift.io
- [x] latest
- [ ] nightly
- [ ] other: please specify
### Steps to reproduce
* Generate ssh keypair *with a password* (`ssh-keygen`)
* Upload public part of that key to gitlab/github
* Start any workspace
* upload private part of the key to the workspace
* try to clone some repo using this format -> `git@gitlab.something/something/web-app.git`
### Expected behavior
Repo is cloned
|
non_process
|
git clone using private key with password is not working describe the bug when trying to clone git repository using git gitlab com something something using ssh key which is protected by password che never asks me for a password and the clone fails with this error git clone git gitlab something something web app git projects web app cloning into projects web app git gitlab com permission denied publickey fatal could not read from remote repository please make sure you have the correct access rights and the repository exists che version che openshift io latest nightly other please specify steps to reproduce generate ssh keypair with a password ssh keygen upload public part of that key to gitlab github start any workspace upload private part of the key to the workspace try to clone some repo using this format git gitlab something something web app git expected behavior repo is cloned
| 0
|
3,925
| 6,845,689,064
|
IssuesEvent
|
2017-11-13 09:17:46
|
openvstorage/volumedriver
|
https://api.github.com/repos/openvstorage/volumedriver
|
closed
|
Redirected volume removal times out and leaves stale metadata
|
priority_normal process_wontfix type_bug
|
Scenario:
- config:
- redirect timeout 5 secs
- backend timeout 15 secs
- volume V is running on node N
- SCO write W to the backend is happening in the background
- volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish
- W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires
=> unlink call times out on N
|
1.0
|
Redirected volume removal times out and leaves stale metadata - Scenario:
- config:
- redirect timeout 5 secs
- backend timeout 15 secs
- volume V is running on node N
- SCO write W to the backend is happening in the background
- volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish
- W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires
=> unlink call times out on N
|
process
|
redirected volume removal times out and leaves stale metadata scenario config redirect timeout secs backend timeout secs volume v is running on node n sco write w to the backend is happening in the background volume removal is started on node m redirect to node n removal needs to wait for w to finish w encounters backend issues finishes unsuccessfully when the backend timeout fires unlink call times out on n
| 1
|
21,434
| 29,477,581,720
|
IssuesEvent
|
2023-06-02 00:34:05
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] [Bug] Incorrect display name for cumulative count operator
|
Type:Bug Querying/GUI Querying/Notebook .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
Once applied, the "Cumulative count" operator is displayed like "Cumulative count of nil" in the notebook editor
This name is coming from `ML.displayInfo(query, index, aggregationClause).displayName`
Can be reproduced on aggregations integration branch (#31209)
### To Reproduce
1. New > Question > Any data source
2. Click "Pick the metric you want to see"
3. Pick the "Cumulative count of rows" option in "Basic operators"
**Expected:** the clause shows "Cumulative count"
**Actual:** the clause shows "Cumulative count of nil"
### Demo
|
1.0
|
[MLv2] [Bug] Incorrect display name for cumulative count operator - Once applied, the "Cumulative count" operator is displayed like "Cumulative count of nil" in the notebook editor
This name is coming from `ML.displayInfo(query, index, aggregationClause).displayName`
Can be reproduced on aggregations integration branch (#31209)
### To Reproduce
1. New > Question > Any data source
2. Click "Pick the metric you want to see"
3. Pick the "Cumulative count of rows" option in "Basic operators"
**Expected:** the clause shows "Cumulative count"
**Actual:** the clause shows "Cumulative count of nil"
### Demo
|
process
|
incorrect display name for cumulative count operator once applied the cumulative count operator is displayed like cumulative count of nil in the notebook editor this name is coming from ml displayinfo query index aggregationclause displayname can be reproduced on aggregations integration branch to reproduce new question any data source click pick the metric you want to see pick the cumulative count of rows option in basic operators expected the clause shows cumulative count actual the clause shows cumulative count of nil demo
| 1
|
16,986
| 22,351,057,948
|
IssuesEvent
|
2022-06-15 12:06:57
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
multiprocessing with maxtasksperchild can hang if unpickling causes import
|
type-bug expert-multiprocessing
|
**Bug report**
This seems like another specific instance of the general issue identified in #50970.
If `multiprocessing.Pool.map_async` is used with `maxtasksperchild` and a value returned by a task is of a class not currently imported by the calling process, it can lead to a hang. Here is an example that reliably hangs for me, but which exits cleanly if ElementTree is imported at the top level.
```python
#!/usr/bin/env python
import os
import multiprocessing
def worker(num: int):
from xml.etree.ElementTree import ElementTree
print(f"Worker {num} with pid {os.getpid()}")
return ElementTree()
def main(cores: int = 4, num: int = 6):
pool = multiprocessing.Pool(processes=cores, maxtasksperchild=1)
barList = list(pool.map_async(worker, list(range(num))).get())
print(barList)
if __name__ == "__main__":
main()
```
Running `py-spy dump` on one of the workers shows this backtrace:
```
Process 47102: python ./demo_core.py
Python v3.10.4 (/usr/bin/python3.10)
Thread 47102 (idle): "Thread-1 (_handle_workers)"
acquire (<frozen importlib._bootstrap>:120)
__enter__ (<frozen importlib._bootstrap>:171)
_find_and_load (<frozen importlib._bootstrap>:1024)
worker (demo_core.py:8)
mapstar (multiprocessing/pool.py:48)
worker (multiprocessing/pool.py:125)
run (multiprocessing/process.py:108)
_bootstrap (multiprocessing/process.py:315)
_launch (multiprocessing/popen_fork.py:71)
__init__ (multiprocessing/popen_fork.py:19)
_Popen (multiprocessing/context.py:277)
start (multiprocessing/process.py:121)
_repopulate_pool_static (multiprocessing/pool.py:326)
_maintain_pool (multiprocessing/pool.py:337)
_handle_workers (multiprocessing/pool.py:513)
run (threading.py:946)
_bootstrap_inner (threading.py:1009)
_bootstrap (threading.py:966)
```
My guess (without any further proof) is that the main process receives a pickled ElementTree and starts importing the module. Concurrently, another thread realises it needs to start a new worker, so does a `fork()`. The child process has a half-imported, locked ElementTree module, and tries to import it again, leading to a deadlock.
Note that this is nothing to do with ElementTree - I get the same behaviour with numpy. I chose ElementTree as a reasonably complex module (to maximise the window for the race condition) with a picklable class.
Personally I consider the `fork` model of `multiprocessing` to be dangerous and requiring of care to ensure all worker tasks are created before doing anything that can conceivably create threads, and definitely a bad combination with `maxtasksperchild`. So I won't shed any tears if the resolution is "won't fix, don't do that". But https://github.com/python/cpython/issues/50970#issuecomment-1147328694 seems to suggest that @vstinner has some appetite for addressing such issues and hence I'm filing this.
**Your environment**
- CPython versions tested on: 3.8.10, 3.10.4
- Operating system and architecture: Ubuntu 20.04, x86_64
|
1.0
|
multiprocessing with maxtasksperchild can hang if unpickling causes import - **Bug report**
This seems like another specific instance of the general issue identified in #50970.
If `multiprocessing.Pool.map_async` is used with `maxtasksperchild` and a value returned by a task is of a class not currently imported by the calling process, it can lead to a hang. Here is an example that reliably hangs for me, but which exits cleanly if ElementTree is imported at the top level.
```python
#!/usr/bin/env python
import os
import multiprocessing
def worker(num: int):
from xml.etree.ElementTree import ElementTree
print(f"Worker {num} with pid {os.getpid()}")
return ElementTree()
def main(cores: int = 4, num: int = 6):
pool = multiprocessing.Pool(processes=cores, maxtasksperchild=1)
barList = list(pool.map_async(worker, list(range(num))).get())
print(barList)
if __name__ == "__main__":
main()
```
Running `py-spy dump` on one of the workers shows this backtrace:
```
Process 47102: python ./demo_core.py
Python v3.10.4 (/usr/bin/python3.10)
Thread 47102 (idle): "Thread-1 (_handle_workers)"
acquire (<frozen importlib._bootstrap>:120)
__enter__ (<frozen importlib._bootstrap>:171)
_find_and_load (<frozen importlib._bootstrap>:1024)
worker (demo_core.py:8)
mapstar (multiprocessing/pool.py:48)
worker (multiprocessing/pool.py:125)
run (multiprocessing/process.py:108)
_bootstrap (multiprocessing/process.py:315)
_launch (multiprocessing/popen_fork.py:71)
__init__ (multiprocessing/popen_fork.py:19)
_Popen (multiprocessing/context.py:277)
start (multiprocessing/process.py:121)
_repopulate_pool_static (multiprocessing/pool.py:326)
_maintain_pool (multiprocessing/pool.py:337)
_handle_workers (multiprocessing/pool.py:513)
run (threading.py:946)
_bootstrap_inner (threading.py:1009)
_bootstrap (threading.py:966)
```
My guess (without any further proof) is that the main process receives a pickled ElementTree and starts importing the module. Concurrently, another thread realises it needs to start a new worker, so does a `fork()`. The child process has a half-imported, locked ElementTree module, and tries to import it again, leading to a deadlock.
Note that this is nothing to do with ElementTree - I get the same behaviour with numpy. I chose ElementTree as a reasonably complex module (to maximise the window for the race condition) with a picklable class.
Personally I consider the `fork` model of `multiprocessing` to be dangerous and requiring of care to ensure all worker tasks are created before doing anything that can conceivably create threads, and definitely a bad combination with `maxtasksperchild`. So I won't shed any tears if the resolution is "won't fix, don't do that". But https://github.com/python/cpython/issues/50970#issuecomment-1147328694 seems to suggest that @vstinner has some appetite for addressing such issues and hence I'm filing this.
**Your environment**
- CPython versions tested on: 3.8.10, 3.10.4
- Operating system and architecture: Ubuntu 20.04, x86_64
|
process
|
multiprocessing with maxtasksperchild can hang if unpickling causes import bug report this seems like another specific instance of the general issue identified in if multiprocessing pool map async is used with maxtasksperchild and a value returned by a task is of a class not currently imported by the calling process it can lead to a hang here is an example that reliably hangs for me but which exits cleanly if elementtree is imported at the top level python usr bin env python import os import multiprocessing def worker num int from xml etree elementtree import elementtree print f worker num with pid os getpid return elementtree def main cores int num int pool multiprocessing pool processes cores maxtasksperchild barlist list pool map async worker list range num get print barlist if name main main running py spy dump on one of the workers shows this backtrace process python demo core py python usr bin thread idle thread handle workers acquire enter find and load worker demo core py mapstar multiprocessing pool py worker multiprocessing pool py run multiprocessing process py bootstrap multiprocessing process py launch multiprocessing popen fork py init multiprocessing popen fork py popen multiprocessing context py start multiprocessing process py repopulate pool static multiprocessing pool py maintain pool multiprocessing pool py handle workers multiprocessing pool py run threading py bootstrap inner threading py bootstrap threading py my guess without any further proof is that the main process receives a pickled elementtree and starts importing the module concurrently another thread realises it needs to start a new worker so does a fork the child process has a half imported locked elementtree module and tries to import it again leading to a deadlock note that this is nothing to do with elementtree i get the same behaviour with numpy i chose elementtree as a reasonably complex module to maximise the window for the race condition with a picklable class personally i consider the fork model of multiprocessing to be dangerous and requiring of care to ensure all worker tasks are created before doing anything that can conceivably create threads and definitely a bad combination with maxtasksperchild so i won t shed any tears if the resolution is won t fix don t do that but seems to suggest that vstinner has some appetite for addressing such issues and hence i m filing this your environment cpython versions tested on operating system and architecture ubuntu
| 1
|
21,598
| 30,002,906,066
|
IssuesEvent
|
2023-06-26 10:28:25
|
ukri-excalibur/excalibur-tests
|
https://api.github.com/repos/ukri-excalibur/excalibur-tests
|
opened
|
CI is failing after ReFrame 4.3.0 release
|
bug postprocessing
|
Our CI runs are failing to find the perflogs, maybe they have changed something in the naming convention.
https://github.com/reframe-hpc/reframe/releases/tag/v4.3.0
|
1.0
|
CI is failing after ReFrame 4.3.0 release - Our CI runs are failing to find the perflogs, maybe they have changed something in the naming convention.
https://github.com/reframe-hpc/reframe/releases/tag/v4.3.0
|
process
|
ci is failing after reframe release our ci runs are failing to find the perflogs maybe they have changed something in the naming convention
| 1
|
118,225
| 25,276,756,702
|
IssuesEvent
|
2022-11-16 13:11:14
|
mozilla/addons-frontend
|
https://api.github.com/repos/mozilla/addons-frontend
|
reopened
|
Revisit the logic around couldShowWarning in InstallWarning
|
component: code quality priority: p3
|
Quoting @willdurand in https://github.com/mozilla/addons-frontend/pull/8348#discussion_r312547172, "I am not sure to like this pattern. Could you file an issue so that we can discuss it and maybe revisit it later?"
|
1.0
|
Revisit the logic around couldShowWarning in InstallWarning - Quoting @willdurand in https://github.com/mozilla/addons-frontend/pull/8348#discussion_r312547172, "I am not sure to like this pattern. Could you file an issue so that we can discuss it and maybe revisit it later?"
|
non_process
|
revisit the logic around couldshowwarning in installwarning quoting willdurand in i am not sure to like this pattern could you file an issue so that we can discuss it and maybe revisit it later
| 0
|
16,625
| 10,545,871,733
|
IssuesEvent
|
2019-10-02 20:08:05
|
cityofaustin/atd-geospatial
|
https://api.github.com/repos/cityofaustin/atd-geospatial
|
closed
|
Upgrade all computers to ArcGIS Desktop 10.6
|
Need: 1-Must Have Service: Geo Type: IT Support Workgroup: DTS
|
I will be working with Scott to upgrade all ATD computers from ArcGIS 10.x to 10.6. Before upgrading need to do the following steps:
- Get a list of all users from Scott
- Email all users notifying them of the upcoming upgrade, asking any users with custom tools to let me know so we can test before upgrading the user and having issues.
- Work with any users that have custom tools to test out in 10.6
- Upgrade the DTS team first so we can ensure there are no issues with the new version before deploying to the entire department.
- Notify users of saving a screenshot of ArcMap or creating a list of all toolbars they normally use and how they like their map document setup. Also what Add-Ins they have installed and use.
- Work with Scott on getting all computers upgraded
- Help users after the upgrade has taken place with any issues, reconnecting of databases, setup of ArcMap, Add-Ins, etc.
|
1.0
|
Upgrade all computers to ArcGIS Desktop 10.6 - I will be working with Scott to upgrade all ATD computers from ArcGIS 10.x to 10.6. Before upgrading need to do the following steps:
- Get a list of all users from Scott
- Email all users notifying them of the upcoming upgrade, asking any users with custom tools to let me know so we can test before upgrading the user and having issues.
- Work with any users that have custom tools to test out in 10.6
- Upgrade the DTS team first so we can ensure there are no issues with the new version before deploying to the entire department.
- Notify users of saving a screenshot of ArcMap or creating a list of all toolbars they normally use and how they like their map document setup. Also what Add-Ins they have installed and use.
- Work with Scott on getting all computers upgraded
- Help users after the upgrade has taken place with any issues, reconnecting of databases, setup of ArcMap, Add-Ins, etc.
|
non_process
|
upgrade all computers to arcgis desktop i will be working with scott to upgrade all atd computers from arcgis x to before upgrading need to do the following steps get a list of all users from scott email all users notifying them of the upcoming upgrade asking any users with custom tools to let me know so we can test before upgrading the user and having issues work with any users that have custom tools to test out in upgrade the dts team first so we can ensure there are no issues with the new version before deploying to the entire department notify users of saving a screenshot of arcmap or creating a list of all toolbars they normally use and how they like their map document setup also what add ins they have installed and use work with scott on getting all computers upgraded help users after the upgrade has taken place with any issues reconnecting of databases setup of arcmap add ins etc
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.