Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
258,136 | 27,563,858,658 | IssuesEvent | 2023-03-08 01:11:32 | LynRodWS/alcor | https://api.github.com/repos/LynRodWS/alcor | opened | CVE-2019-20330 (High) detected in jackson-databind-2.9.9.jar | security vulnerability | ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /services/api_gateway/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-netflix-hystrix-2.1.2.RELEASE.jar (Root Library)
- hystrix-serialization-1.5.18.jar
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.9.10.2</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-netflix-hystrix): 2.1.3.RELEASE</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2019-20330 (High) detected in jackson-databind-2.9.9.jar - ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /services/api_gateway/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-netflix-hystrix-2.1.2.RELEASE.jar (Root Library)
- hystrix-serialization-1.5.18.jar
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.9.10.2</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-netflix-hystrix): 2.1.3.RELEASE</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file services api gateway pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring cloud starter netflix hystrix release jar root library hystrix serialization jar x jackson databind jar vulnerable library found in base branch master vulnerability details fasterxml jackson databind x before lacks certain net sf ehcache blocking publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org springframework cloud spring cloud starter netflix hystrix release rescue worker helmet automatic remediation is available for this issue | 0 |
39,428 | 16,012,930,629 | IssuesEvent | 2021-04-20 12:57:15 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | [URGENT] - Resolve issue related to VZA login | Impact: 1-Severe Product: Vision Zero in Action Service: Apps Type: Bug Report Workgroup: VZ | <!-- Email -->
<!-- Ramon.Sanchez@austintexas.gov -->
> What application are you using?
Vision Zero (Viewer)
> Describe the problem.
(Safari) Can’t log in. I click sign in and the page just refreshes to the same page. I see the link address change at the top, but it reverts to the original log in page.
(Chrome) Can log in, but it just reverts back to the log in page. Never able to actually get into the application.
> Is there anything else we should know?
APD trying to sign up for shifts
> Requested By
Ramon S.
Request ID: DTS21-101565
| 1.0 | [URGENT] - Resolve issue related to VZA login - <!-- Email -->
<!-- Ramon.Sanchez@austintexas.gov -->
> What application are you using?
Vision Zero (Viewer)
> Describe the problem.
(Safari) Can’t log in. I click sign in and the page just refreshes to the same page. I see the link address change at the top, but it reverts to the original log in page.
(Chrome) Can log in, but it just reverts back to the log in page. Never able to actually get into the application.
> Is there anything else we should know?
APD trying to sign up for shifts
> Requested By
Ramon S.
Request ID: DTS21-101565
| non_priority | resolve issue related to vza login what application are you using vision zero viewer describe the problem safari can’t log in i click sign in and the page just refreshes to the same page i see the link address change at the top but it reverts to the original log in page chrome can log in but it just reverts back to the log in page never able to actually get into the application is there anything else we should know apd trying to sign up for shifts requested by ramon s request id | 0 |
17,567 | 3,621,473,842 | IssuesEvent | 2016-02-09 00:21:29 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | UDP tests fail with timeout sending/receiving data over loopback | 2 - In Progress System.Net test bug | http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/windows_nt_release_prtest/956/consoleFull
```
17:49:39 System.Net.Sockets.Tests.SendReceive.SendToRecvFromAsync_UdpClient_Single_Datagram_UDP_IPv4 [FAIL]
17:49:39 Assert.True() Failure
17:49:39 Expected: True
17:49:39 Actual: False
17:49:39 Stack Trace:
17:49:39 d:\j\workspace\dotnet_corefx\windows_nt_release_prtest\src\System.Net.Sockets\tests\FunctionalTests\SendReceive.cs(192,0): at System.Net.Sockets.Tests.SendReceive.SendToRecvFromAsync_UdpClient_Datagram_UDP(IPAddress leftAddress, IPAddress rightAddress)
``` | 1.0 | UDP tests fail with timeout sending/receiving data over loopback - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/windows_nt_release_prtest/956/consoleFull
```
17:49:39 System.Net.Sockets.Tests.SendReceive.SendToRecvFromAsync_UdpClient_Single_Datagram_UDP_IPv4 [FAIL]
17:49:39 Assert.True() Failure
17:49:39 Expected: True
17:49:39 Actual: False
17:49:39 Stack Trace:
17:49:39 d:\j\workspace\dotnet_corefx\windows_nt_release_prtest\src\System.Net.Sockets\tests\FunctionalTests\SendReceive.cs(192,0): at System.Net.Sockets.Tests.SendReceive.SendToRecvFromAsync_UdpClient_Datagram_UDP(IPAddress leftAddress, IPAddress rightAddress)
``` | non_priority | udp tests fail with timeout sending receiving data over loopback system net sockets tests sendreceive sendtorecvfromasync udpclient single datagram udp assert true failure expected true actual false stack trace d j workspace dotnet corefx windows nt release prtest src system net sockets tests functionaltests sendreceive cs at system net sockets tests sendreceive sendtorecvfromasync udpclient datagram udp ipaddress leftaddress ipaddress rightaddress | 0 |
6,702 | 3,439,516,515 | IssuesEvent | 2015-12-14 09:59:39 | agdsn/sipa | https://api.github.com/repos/agdsn/sipa | opened | Make decision of uwsgi usage easier | code enhancement | Currently, if one wants to use uwsgi or not, has some significant consequences regarding how sipa has to be started. To make this easier, there should be a start script, which decides based on an environment variable like `SIPA_USE_UWSGI`, what is run. Also, this “decision” could be passed to sipa to make some things easier, e.g. there does not have to be the monkey-patch when using uwsgi (or the other way around, I always forget).
This could perhaps be combined with the prefix, so keep #176 in mind. | 1.0 | Make decision of uwsgi usage easier - Currently, if one wants to use uwsgi or not, has some significant consequences regarding how sipa has to be started. To make this easier, there should be a start script, which decides based on an environment variable like `SIPA_USE_UWSGI`, what is run. Also, this “decision” could be passed to sipa to make some things easier, e.g. there does not have to be the monkey-patch when using uwsgi (or the other way around, I always forget).
This could perhaps be combined with the prefix, so keep #176 in mind. | non_priority | make decision of uwsgi usage easier currently if one wants to use uwsgi or not has some significant consequences regarding how sipa has to be started to make this easier there should be a start script which decides based on an environment variable like sipa use uwsgi what is run also this “decision” could be passed to sipa to make some things easier e g there does not have to be the monkey patch when using uwsgi or the other way around i always forget this could perhaps be combined with the prefix so keep in mind | 0 |
211,615 | 16,329,720,729 | IssuesEvent | 2021-05-12 07:42:02 | ckeditor/ckeditor4 | https://api.github.com/repos/ckeditor/ckeditor4 | opened | Failing test: /tests/plugins/pagebreak/pagebreak | browser:safari status:confirmed type:failingtest | ## Type of report
Failing test
## Provide detailed reproduction steps (if any)
The test included in `/tests/plugins/pagebreak/pagebreak` fails
Name of test: [test paste](tests/plugins/pagebreak/pagebreak#tests%2Fplugins%2Fpagebreak%2Fpagebreak%20test%20paste) <br>
Errors: <br>

### Expected result
* Test passed
### Actual result
* Test failed
## Other details
* Browser: Safari
* OS: macOS
* CKEditor version: 4.16.1
| 1.0 | Failing test: /tests/plugins/pagebreak/pagebreak - ## Type of report
Failing test
## Provide detailed reproduction steps (if any)
The test included in `/tests/plugins/pagebreak/pagebreak` fails
Name of test: [test paste](tests/plugins/pagebreak/pagebreak#tests%2Fplugins%2Fpagebreak%2Fpagebreak%20test%20paste) <br>
Errors: <br>

### Expected result
* Test passed
### Actual result
* Test failed
## Other details
* Browser: Safari
* OS: macOS
* CKEditor version: 4.16.1
| non_priority | failing test tests plugins pagebreak pagebreak type of report failing test provide detailed reproduction steps if any the test included in tests plugins pagebreak pagebreak fails name of test tests plugins pagebreak pagebreak tests errors expected result test passed actual result test failed other details browser safari os macos ckeditor version | 0 |
6,832 | 3,059,953,265 | IssuesEvent | 2015-08-14 17:51:47 | bigdatagenomics/adam | https://api.github.com/repos/bigdatagenomics/adam | opened | Updating/Publishing the docs/ directory | discussion documentation question | I only realized that the `docs/` directory exists recently, when refactoring README.md (#763, #764).
@fnothaft [implied on #764](https://github.com/bigdatagenomics/adam/pull/764#issuecomment-130507272) that the docs there get distributed via released artifacts, but I don't see any in [adam-distribution_2.10-0.17.0-bin.tar.gz](https://repo1.maven.org/maven2/org/bdgenomics/adam/adam-distribution_2.10/0.17.0/adam-distribution_2.10-0.17.0-bin.tar.gz).
It seems like there was a lot of somewhat-duplicated markdown living there, vs. what was in README; now the README is very different.
Are builds of `docs/` available on the internet anywhere? How should we keep them updated / consistent with README.md? Is it up to anyone making changes to update the docs in both places?
Should we move towards setting up e.g. [readthedocs.org](https://readthedocs.org/) builds of the docs for each version? Should the README point at those? I generally like that approach, though some way to also publish docs for random non-release SHAs/nightly-snapshots or something would also be good and has been lacking in other projects who've done something like this, e.g. Spark. | 1.0 | Updating/Publishing the docs/ directory - I only realized that the `docs/` directory exists recently, when refactoring README.md (#763, #764).
@fnothaft [implied on #764](https://github.com/bigdatagenomics/adam/pull/764#issuecomment-130507272) that the docs there get distributed via released artifacts, but I don't see any in [adam-distribution_2.10-0.17.0-bin.tar.gz](https://repo1.maven.org/maven2/org/bdgenomics/adam/adam-distribution_2.10/0.17.0/adam-distribution_2.10-0.17.0-bin.tar.gz).
It seems like there was a lot of somewhat-duplicated markdown living there, vs. what was in README; now the README is very different.
Are builds of `docs/` available on the internet anywhere? How should we keep them updated / consistent with README.md? Is it up to anyone making changes to update the docs in both places?
Should we move towards setting up e.g. [readthedocs.org](https://readthedocs.org/) builds of the docs for each version? Should the README point at those? I generally like that approach, though some way to also publish docs for random non-release SHAs/nightly-snapshots or something would also be good and has been lacking in other projects who've done something like this, e.g. Spark. | non_priority | updating publishing the docs directory i only realized that the docs directory exists recently when refactoring readme md fnothaft that the docs there get distributed via released artifacts but i don t see any in it seems like there was a lot of somewhat duplicated markdown living there vs what was in readme now the readme is very different are builds of docs available on the internet anywhere how should we keep them updated consistent with readme md is it up to anyone making changes to update the docs in both places should we move towards setting up e g builds of the docs for each version should the readme point at those i generally like that approach though some way to also publish docs for random non release shas nightly snapshots or something would also be good and has been lacking in other projects who ve done something like this e g spark | 0 |
296,611 | 22,309,479,905 | IssuesEvent | 2022-06-13 15:40:49 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | OS.move_to_trash() not functioning or poorly documented | topic:core documentation | ### Godot version
4.0 Alpha 9
### System information
Windows 10
### Issue description
Using OS.move_to_trash always produces a generic failure (error code 1). This could be due to my windows configuration, or it could also be due to a bug, or it could be because OS.move_to_trash() is very minimally documented.
### Steps to reproduce
Run the project and click the button.
```
func _on_button_pressed():
var result = OS.move_to_trash("res://test.txt")
print(result)
```
### Minimal reproduction project
[Move_To_TrashProject.zip](https://github.com/godotengine/godot/files/8884618/Move_To_TrashProject.zip)
| 1.0 | OS.move_to_trash() not functioning or poorly documented - ### Godot version
4.0 Alpha 9
### System information
Windows 10
### Issue description
Using OS.move_to_trash always produces a generic failure (error code 1). This could be due to my windows configuration, or it could also be due to a bug, or it could be because OS.move_to_trash() is very minimally documented.
### Steps to reproduce
Run the project and click the button.
```
func _on_button_pressed():
var result = OS.move_to_trash("res://test.txt")
print(result)
```
### Minimal reproduction project
[Move_To_TrashProject.zip](https://github.com/godotengine/godot/files/8884618/Move_To_TrashProject.zip)
| non_priority | os move to trash not functioning or poorly documented godot version alpha system information windows issue description using os move to trash always produces a generic failure error code this could be due to my windows configuration or it could also be due to a bug or it could be because os move to trash is very minimally documented steps to reproduce run the project and click the button func on button pressed var result os move to trash res test txt print result minimal reproduction project | 0 |
450,212 | 31,885,482,178 | IssuesEvent | 2023-09-16 22:23:45 | sparcs-kaist/zabo-front-reactjs | https://api.github.com/repos/sparcs-kaist/zabo-front-reactjs | closed | GitHub Actions, Issue Template에서 오타 제거 | documentation | # Summary <!-- 간단한 요약. -->
GitHub Actions, Issue Template에서 오타를 제거합니다.
# Tasks <!-- 해야 할 태스크 체크리스트로 만들기 -->
- [ ] 오타 제거
| 1.0 | GitHub Actions, Issue Template에서 오타 제거 - # Summary <!-- 간단한 요약. -->
GitHub Actions, Issue Template에서 오타를 제거합니다.
# Tasks <!-- 해야 할 태스크 체크리스트로 만들기 -->
- [ ] 오타 제거
| non_priority | github actions issue template에서 오타 제거 summary github actions issue template에서 오타를 제거합니다 tasks 오타 제거 | 0 |
124,840 | 26,547,528,941 | IssuesEvent | 2023-01-20 02:34:38 | geostyler/geostyler | https://api.github.com/repos/geostyler/geostyler | closed | Create geostyler-geopackage-parser | feature complex parser-request codesprint stale | Create the `geostyler-geopackage-parser`.
Implementation of the GeoStyler_DataParser interface targeting [geopackage](https://de.wikipedia.org/wiki/GeoPackage) based geodata.
| 1.0 | Create geostyler-geopackage-parser - Create the `geostyler-geopackage-parser`.
Implementation of the GeoStyler_DataParser interface targeting [geopackage](https://de.wikipedia.org/wiki/GeoPackage) based geodata.
| non_priority | create geostyler geopackage parser create the geostyler geopackage parser implementation of the geostyler dataparser interface targeting based geodata | 0 |
249,877 | 21,213,406,377 | IssuesEvent | 2022-04-11 03:31:35 | metaplex-foundation/metaplex | https://api.github.com/repos/metaplex-foundation/metaplex | closed | End sale & claim item doesn't work | needs tests Store Front bug Stale | ### Which package is this bug report for?
Metaplex Storefront
### Issue description
When I go to the live section, and I want to remove a nft listed for sale, and I click on end sale & claim items nothing happens and it does not make me close the auction and collect the nft. Does the same thing happen to anyone? i don't know how to remove the rod and collect the nft.
Thanks a lot and anyone helping me to fix this.
| 1.0 | End sale & claim item doesn't work - ### Which package is this bug report for?
Metaplex Storefront
### Issue description
When I go to the live section, and I want to remove a nft listed for sale, and I click on end sale & claim items nothing happens and it does not make me close the auction and collect the nft. Does the same thing happen to anyone? i don't know how to remove the rod and collect the nft.
Thanks a lot and anyone helping me to fix this.
| non_priority | end sale claim item doesn t work which package is this bug report for metaplex storefront issue description when i go to the live section and i want to remove a nft listed for sale and i click on end sale claim items nothing happens and it does not make me close the auction and collect the nft does the same thing happen to anyone i don t know how to remove the rod and collect the nft thanks a lot and anyone helping me to fix this | 0 |
67,314 | 27,794,639,690 | IssuesEvent | 2023-03-17 11:33:15 | gaelreyrol/nixos-config | https://api.github.com/repos/gaelreyrol/nixos-config | opened | Configure Home Assistant on APU | enhancement services lab | With:
- PostgreSQL database
- Backups to Clever Cloud Cellar | 1.0 | Configure Home Assistant on APU - With:
- PostgreSQL database
- Backups to Clever Cloud Cellar | non_priority | configure home assistant on apu with postgresql database backups to clever cloud cellar | 0 |
59,842 | 14,663,266,371 | IssuesEvent | 2020-12-29 09:19:06 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | opened | Error when build TF 1.13.2 from the source | type:build/install | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution: Linux Ubuntu 16.04/18.04
- TensorFlow installed from source
- TensorFlow version: 1.13.2
- Python version: 2.7
- Installed using: docker
- Bazel version: 0.19.2
- GCC/Compiler version: 4.9
- CUDA/cuDNN version: 10.0, 7
- GPU model and memory: T4, 32GB
**Describe the problem**
I tried to install TF from the source following instructions here, https://github.com/tensorflow/docs/blob/r1.13/site/en/install/source.md#build-from-source.
However some error occurs: ERROR: /home/cluster/tensorflow/tensorflow/core/kernels/BUILD:593:1: C++ compilation of rule '//tensorflow/core/kernels:eigen_contraction_kernel' failed (Exit 1). Any suggestions?
| 1.0 | Error when build TF 1.13.2 from the source - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution: Linux Ubuntu 16.04/18.04
- TensorFlow installed from source
- TensorFlow version: 1.13.2
- Python version: 2.7
- Installed using: docker
- Bazel version: 0.19.2
- GCC/Compiler version: 4.9
- CUDA/cuDNN version: 10.0, 7
- GPU model and memory: T4, 32GB
**Describe the problem**
I tried to install TF from the source following instructions here, https://github.com/tensorflow/docs/blob/r1.13/site/en/install/source.md#build-from-source.
However some error occurs: ERROR: /home/cluster/tensorflow/tensorflow/core/kernels/BUILD:593:1: C++ compilation of rule '//tensorflow/core/kernels:eigen_contraction_kernel' failed (Exit 1). Any suggestions?
| non_priority | error when build tf from the source please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution linux ubuntu tensorflow installed from source tensorflow version python version installed using docker bazel version gcc compiler version cuda cudnn version gpu model and memory describe the problem i tried to install tf from the source following instructions here however some error occurs error home cluster tensorflow tensorflow core kernels build c compilation of rule tensorflow core kernels eigen contraction kernel failed exit any suggestions | 0 |
239,611 | 19,904,036,170 | IssuesEvent | 2022-01-25 10:53:14 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] ConcurrentSnapshotsIT testMasterFailoverDuringStaleIndicesCleanup failing | :Distributed/Snapshot/Restore >test-failure Team:Distributed | **Build scan:**
https://gradle-enterprise.elastic.co/s/kghqtdl7uxnz6/tests/:server:internalClusterTest/org.elasticsearch.snapshots.ConcurrentSnapshotsIT/testMasterFailoverDuringStaleIndicesCleanup
**Reproduction line:**
`./gradlew ':server:internalClusterTest' --tests "org.elasticsearch.snapshots.ConcurrentSnapshotsIT.testMasterFailoverDuringStaleIndicesCleanup" -Dtests.seed=3A1F7F594461254 -Dtests.locale=en-NZ -Dtests.timezone=CAT -Druntime.java=8`
**Applicable branches:**
7.17
**Reproduces locally?:**
Yes
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.snapshots.ConcurrentSnapshotsIT&tests.test=testMasterFailoverDuringStaleIndicesCleanup
**Failure excerpt:**
```
java.lang.AssertionError: (No message provided)
at __randomizedtesting.SeedInfo.seed([3A1F7F594461254:3C24A6936A33287]:0)
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at org.elasticsearch.snapshots.AbstractSnapshotIntegTestCase.lambda$waitForBlock$5(AbstractSnapshotIntegTestCase.java:212)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1140)
at org.elasticsearch.snapshots.AbstractSnapshotIntegTestCase.waitForBlock(AbstractSnapshotIntegTestCase.java:212)
at org.elasticsearch.snapshots.ConcurrentSnapshotsIT.testMasterFailoverDuringStaleIndicesCleanup(ConcurrentSnapshotsIT.java:1086)
at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
``` | 1.0 | [CI] ConcurrentSnapshotsIT testMasterFailoverDuringStaleIndicesCleanup failing - **Build scan:**
https://gradle-enterprise.elastic.co/s/kghqtdl7uxnz6/tests/:server:internalClusterTest/org.elasticsearch.snapshots.ConcurrentSnapshotsIT/testMasterFailoverDuringStaleIndicesCleanup
**Reproduction line:**
`./gradlew ':server:internalClusterTest' --tests "org.elasticsearch.snapshots.ConcurrentSnapshotsIT.testMasterFailoverDuringStaleIndicesCleanup" -Dtests.seed=3A1F7F594461254 -Dtests.locale=en-NZ -Dtests.timezone=CAT -Druntime.java=8`
**Applicable branches:**
7.17
**Reproduces locally?:**
Yes
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.snapshots.ConcurrentSnapshotsIT&tests.test=testMasterFailoverDuringStaleIndicesCleanup
**Failure excerpt:**
```
java.lang.AssertionError: (No message provided)
at __randomizedtesting.SeedInfo.seed([3A1F7F594461254:3C24A6936A33287]:0)
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at org.elasticsearch.snapshots.AbstractSnapshotIntegTestCase.lambda$waitForBlock$5(AbstractSnapshotIntegTestCase.java:212)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1140)
at org.elasticsearch.snapshots.AbstractSnapshotIntegTestCase.waitForBlock(AbstractSnapshotIntegTestCase.java:212)
at org.elasticsearch.snapshots.ConcurrentSnapshotsIT.testMasterFailoverDuringStaleIndicesCleanup(ConcurrentSnapshotsIT.java:1086)
at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
``` | non_priority | concurrentsnapshotsit testmasterfailoverduringstaleindicescleanup failing build scan reproduction line gradlew server internalclustertest tests org elasticsearch snapshots concurrentsnapshotsit testmasterfailoverduringstaleindicescleanup dtests seed dtests locale en nz dtests timezone cat druntime java applicable branches reproduces locally yes failure history failure excerpt java lang assertionerror no message provided at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert asserttrue assert java at org elasticsearch snapshots abstractsnapshotintegtestcase lambda waitforblock abstractsnapshotintegtestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch snapshots abstractsnapshotintegtestcase waitforblock abstractsnapshotintegtestcase java at org elasticsearch snapshots concurrentsnapshotsit testmasterfailoverduringstaleindicescleanup concurrentsnapshotsit java at sun reflect nativemethodaccessorimpl nativemethodaccessorimpl java at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at java lang thread run thread java | 0 |
650 | 2,594,481,851 | IssuesEvent | 2015-02-20 04:04:45 | BALL-Project/ball | https://api.github.com/repos/BALL-Project/ball | closed | std::string::pop_back not supported on gcc 4.6.3 even with -std=c++0x enabled | C: Buildsystem P: blocker R: fixed T: defect | **Reported by nicste on 3 Jun 44324721 15:40 UTC**
Hey everyone,
gcc/g++ 4.6.3, the shipped compiler with (K)Ubuntu 12.04.x, the previous and still widely used LTS version, does not support string::pop_back as introduced with c++11. Even with -std=c++0x it does not work, support was introduced with compiler version 4.7.
/usr/bin/g++ -DBALL_EXPORTS -DQT_CORE_LIB -DQT_GUI_LIB -DQT_NETWORK_LIB -DQT_NO_DEBUG -DQT_OPENGL_LIB -DQT_SQL_LIB -DQT_TEST_LIB -DQT_WEBKIT_LIB -DQT_XML_LIB -O3 -DNDEBUG -fPIC -I/home/AH/nicste/BALL.cmake/build/include -I/home/AH/nicste/BALL.cmake/include -I/usr/lib/openmpi/include -I/usr/lib/openmpi/include/openmpi -isystem /usr/include/qt4 -isystem /usr/include/qt4/QtOpenGL -isystem /usr/include/qt4/QtWebKit -isystem /usr/include/qt4/QtGui -isystem /usr/include/qt4/QtTest -isystem /usr/include/qt4/QtXml -isystem /usr/include/qt4/QtSql -isystem /usr/include/qt4/QtNetwork -isystem /usr/include/qt4/QtCore -I/usr/include/eigen3 -I/usr/include/python2.7 -I/usr/include/openbabel-2.0 -std=c++0x -Wall -Wextra -Wno-long-long -pedantic -pedantic -std=c++0x -o CMakeFiles/BALL.dir/source/CONCEPT/composite.o -c /home/AH/nicste/BALL.cmake/source/CONCEPT/composite.C
In file included from /home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/string.h:1895:0,
from /home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/hashMap.h:13,
from /home/AH/nicste/BALL.cmake/include/BALL/CONCEPT/persistenceManager.h:13,
from /home/AH/nicste/BALL.cmake/include/BALL/CONCEPT/timeStamp.h:21,
from /home/AH/nicste/BALL.cmake/include/BALL/CONCEPT/composite.h:41,
from /home/AH/nicste/BALL.cmake/source/CONCEPT/composite.C:7:
/home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/string.iC: In member function 'void BALL::String::pop_back()':
/home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/string.iC:1545:7: error: 'std::string' has no member named 'pop_back'
make[*** [CMakeFiles/BALL.dir/source/CONCEPT/composite.o](2]:) Error 1
make[Leaving directory `/home/AH/nicste/BALL.cmake/build'
make[1](2]:): *** [Error 2
make[1](CMakeFiles/BALL.dir/all]): Leaving directory `/home/AH/nicste/BALL.cmake/build'
make: *** [all] Error 2
| 1.0 | std::string::pop_back not supported on gcc 4.6.3 even with -std=c++0x enabled - **Reported by nicste on 3 Jun 44324721 15:40 UTC**
Hey everyone,
gcc/g++ 4.6.3, the shipped compiler with (K)Ubuntu 12.04.x, the previous and still widely used LTS version, does not support string::pop_back as introduced with c++11. Even with -std=c++0x it does not work, support was introduced with compiler version 4.7.
/usr/bin/g++ -DBALL_EXPORTS -DQT_CORE_LIB -DQT_GUI_LIB -DQT_NETWORK_LIB -DQT_NO_DEBUG -DQT_OPENGL_LIB -DQT_SQL_LIB -DQT_TEST_LIB -DQT_WEBKIT_LIB -DQT_XML_LIB -O3 -DNDEBUG -fPIC -I/home/AH/nicste/BALL.cmake/build/include -I/home/AH/nicste/BALL.cmake/include -I/usr/lib/openmpi/include -I/usr/lib/openmpi/include/openmpi -isystem /usr/include/qt4 -isystem /usr/include/qt4/QtOpenGL -isystem /usr/include/qt4/QtWebKit -isystem /usr/include/qt4/QtGui -isystem /usr/include/qt4/QtTest -isystem /usr/include/qt4/QtXml -isystem /usr/include/qt4/QtSql -isystem /usr/include/qt4/QtNetwork -isystem /usr/include/qt4/QtCore -I/usr/include/eigen3 -I/usr/include/python2.7 -I/usr/include/openbabel-2.0 -std=c++0x -Wall -Wextra -Wno-long-long -pedantic -pedantic -std=c++0x -o CMakeFiles/BALL.dir/source/CONCEPT/composite.o -c /home/AH/nicste/BALL.cmake/source/CONCEPT/composite.C
In file included from /home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/string.h:1895:0,
from /home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/hashMap.h:13,
from /home/AH/nicste/BALL.cmake/include/BALL/CONCEPT/persistenceManager.h:13,
from /home/AH/nicste/BALL.cmake/include/BALL/CONCEPT/timeStamp.h:21,
from /home/AH/nicste/BALL.cmake/include/BALL/CONCEPT/composite.h:41,
from /home/AH/nicste/BALL.cmake/source/CONCEPT/composite.C:7:
/home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/string.iC: In member function 'void BALL::String::pop_back()':
/home/AH/nicste/BALL.cmake/include/BALL/DATATYPE/string.iC:1545:7: error: 'std::string' has no member named 'pop_back'
make[*** [CMakeFiles/BALL.dir/source/CONCEPT/composite.o](2]:) Error 1
make[Leaving directory `/home/AH/nicste/BALL.cmake/build'
make[1](2]:): *** [Error 2
make[1](CMakeFiles/BALL.dir/all]): Leaving directory `/home/AH/nicste/BALL.cmake/build'
make: *** [all] Error 2
| non_priority | std string pop back not supported on gcc even with std c enabled reported by nicste on jun utc hey everyone gcc g the shipped compiler with k ubuntu x the previous and still widely used lts version does not support string pop back as introduced with c even with std c it does not work support was introduced with compiler version usr bin g dball exports dqt core lib dqt gui lib dqt network lib dqt no debug dqt opengl lib dqt sql lib dqt test lib dqt webkit lib dqt xml lib dndebug fpic i home ah nicste ball cmake build include i home ah nicste ball cmake include i usr lib openmpi include i usr lib openmpi include openmpi isystem usr include isystem usr include qtopengl isystem usr include qtwebkit isystem usr include qtgui isystem usr include qttest isystem usr include qtxml isystem usr include qtsql isystem usr include qtnetwork isystem usr include qtcore i usr include i usr include i usr include openbabel std c wall wextra wno long long pedantic pedantic std c o cmakefiles ball dir source concept composite o c home ah nicste ball cmake source concept composite c in file included from home ah nicste ball cmake include ball datatype string h from home ah nicste ball cmake include ball datatype hashmap h from home ah nicste ball cmake include ball concept persistencemanager h from home ah nicste ball cmake include ball concept timestamp h from home ah nicste ball cmake include ball concept composite h from home ah nicste ball cmake source concept composite c home ah nicste ball cmake include ball datatype string ic in member function void ball string pop back home ah nicste ball cmake include ball datatype string ic error std string has no member named pop back make error make leaving directory home ah nicste ball cmake build make error make cmakefiles ball dir all leaving directory home ah nicste ball cmake build make error | 0 |
76,155 | 26,269,656,143 | IssuesEvent | 2023-01-06 15:47:40 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | closed | voice broadcast:broadcasting event still on the screen after removing... | T-Defect S-Minor O-Uncommon A-Broadcast | ### Steps to reproduce
1. when sending voice, no visualizer...
2. fail to remove broadcasting even deleted on the desktop
### Outcome
just wonder why this happened on the mobile
### Your phone model
_No response_
### Operating system version
android 12
### Application version and app store
1.54
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No | 1.0 | voice broadcast:broadcasting event still on the screen after removing... - ### Steps to reproduce
1. when sending voice, no visualizer...
2. fail to remove broadcasting even deleted on the desktop
### Outcome
just wonder why this happened on the mobile
### Your phone model
_No response_
### Operating system version
android 12
### Application version and app store
1.54
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No | non_priority | voice broadcast:broadcasting event still on the screen after removing steps to reproduce when sending voice no visualizer fail to remove broadcasting even deleted on the desktop outcome just wonder why this happened on the mobile your phone model no response operating system version android application version and app store homeserver no response will you send logs no are you willing to provide a pr no | 0 |
186,454 | 14,394,699,409 | IssuesEvent | 2020-12-03 01:55:09 | github-vet/rangeclosure-findings | https://api.github.com/repos/github-vet/rangeclosure-findings | closed | zhuyawen/biscuit: src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go; 14 LoC | fresh small test |
Found a possible issue in [zhuyawen/biscuit](https://www.github.com/zhuyawen/biscuit) at [src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go](https://github.com/zhuyawen/biscuit/blob/70d8496d1d31ec332f09200184229b743d433d45/src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go#L236-L249)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/zhuyawen/biscuit/blob/70d8496d1d31ec332f09200184229b743d433d45/src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go#L236-L249)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for i, want := range wants {
b, err := want.Pack()
if err != nil {
t.Fatalf("%d: packing failed: %v", i, err)
}
var got Message
err = got.Unpack(b)
if err != nil {
t.Fatalf("%d: unpacking failed: %v", i, err)
}
if !reflect.DeepEqual(got, want) {
t.Errorf("%d: got = %+v, want = %+v", i, &got, &want)
}
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to want at line 247 may start a goroutine
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 70d8496d1d31ec332f09200184229b743d433d45
| 1.0 | zhuyawen/biscuit: src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go; 14 LoC -
Found a possible issue in [zhuyawen/biscuit](https://www.github.com/zhuyawen/biscuit) at [src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go](https://github.com/zhuyawen/biscuit/blob/70d8496d1d31ec332f09200184229b743d433d45/src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go#L236-L249)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/zhuyawen/biscuit/blob/70d8496d1d31ec332f09200184229b743d433d45/src/vendor/golang_org/x/net/dns/dnsmessage/message_test.go#L236-L249)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for i, want := range wants {
b, err := want.Pack()
if err != nil {
t.Fatalf("%d: packing failed: %v", i, err)
}
var got Message
err = got.Unpack(b)
if err != nil {
t.Fatalf("%d: unpacking failed: %v", i, err)
}
if !reflect.DeepEqual(got, want) {
t.Errorf("%d: got = %+v, want = %+v", i, &got, &want)
}
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to want at line 247 may start a goroutine
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 70d8496d1d31ec332f09200184229b743d433d45
| non_priority | zhuyawen biscuit src vendor golang org x net dns dnsmessage message test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for i want range wants b err want pack if err nil t fatalf d packing failed v i err var got message err got unpack b if err nil t fatalf d unpacking failed v i err if reflect deepequal got want t errorf d got v want v i got want below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to want at line may start a goroutine leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
87,580 | 8,101,312,440 | IssuesEvent | 2018-08-12 12:16:05 | curationexperts/laevigata | https://api.github.com/repos/curationexperts/laevigata | closed | Allow Approvers to manage embargos for their school post graduation | acceptance testing embargo review | - [ ] As an approver (not an admin) I can edit the embargo after the student has graduated
- [ ] How do we demo this? | 1.0 | Allow Approvers to manage embargos for their school post graduation - - [ ] As an approver (not an admin) I can edit the embargo after the student has graduated
- [ ] How do we demo this? | non_priority | allow approvers to manage embargos for their school post graduation as an approver not an admin i can edit the embargo after the student has graduated how do we demo this | 0 |
9,679 | 30,230,920,653 | IssuesEvent | 2023-07-06 06:50:51 | z0ph/status | https://api.github.com/repos/z0ph/status | closed | 🛑 Home Automation is down | status home-automation | In [`95fd57c`](https://github.com/z0ph/status/commit/95fd57c0aec45374834a65885662185dbb5b3bea
), Home Automation ($SECRET_SITE) was **down**:
- HTTP code: 0
- Response time: 0 ms
| 1.0 | 🛑 Home Automation is down - In [`95fd57c`](https://github.com/z0ph/status/commit/95fd57c0aec45374834a65885662185dbb5b3bea
), Home Automation ($SECRET_SITE) was **down**:
- HTTP code: 0
- Response time: 0 ms
| non_priority | 🛑 home automation is down in home automation secret site was down http code response time ms | 0 |
223,368 | 24,711,789,362 | IssuesEvent | 2022-10-20 01:48:28 | AlexRogalskiy/github-action-node-dependency | https://api.github.com/repos/AlexRogalskiy/github-action-node-dependency | opened | CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz | security vulnerability | ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>
Dependency Hierarchy:
- eslint-7.24.0.tgz (Root Library)
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-node-dependency/commit/7557b6be7ecb09025024a5fb56ec93d5aba43c03">7557b6be7ecb09025024a5fb56ec93d5aba43c03</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz - ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>
Dependency Hierarchy:
- eslint-7.24.0.tgz (Root Library)
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-node-dependency/commit/7557b6be7ecb09025024a5fb56ec93d5aba43c03">7557b6be7ecb09025024a5fb56ec93d5aba43c03</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in minimatch tgz cve high severity vulnerability vulnerable library minimatch tgz a glob matcher in javascript library home page a href dependency hierarchy eslint tgz root library x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch step up your open source security game with mend | 0 |
6,598 | 5,541,289,145 | IssuesEvent | 2017-03-22 12:27:36 | raoulvdberge/refinedstorage | https://api.github.com/repos/raoulvdberge/refinedstorage | closed | Extreme server lag with updating grid with new version of refinedstorage | investigate performance | Hello. I just upgraded refinedstorage on my server from version 1.2.16 to 1.2.24. However, in this version of refinedstorage, whenever I open any item grid: The grid opens instantly, but it does not show up to date items, The server freezes for 2-10 SECONDS, then the grid shows the items. There was no noticable lag in the older version of refinedstorage.
I was using ftb hermitpack with a couple of mods upgraded/added. (For a full list of changed mods, please refer to [this pastebin](http://pastebin.com/CvSQ2qqy) )
I have a midtier storage system with:
~60 crafters for crafting with recipies (I have to admit that many of them have ore dictionary, I was not aware of the performance consequences when I made them)
~20 crafters for machines
~20 external storage deep storage units
-3 disk drives
-The system is split over 2 remote locations in the same dimensions
The lag does not occur with fluid grids.
**I want to stress that I did not experience any lag in version 1.2.16 when opening the grid.**
I took some profiles of the lag:
[https://transfer.sh/EmO8n/refinedstorage.nps](https://transfer.sh/EmO8n/refinedstorage.nps)
I started recording, opened the crafting grid once, used /forge tps a couple times to see if the lag was gone, stopped recording
[https://transfer.sh/ZWqzQ/refinedstorage.nps](https://transfer.sh/ZWqzQ/refinedstorage.nps)
I started profiling, opened the crafting grid twice, used /forge tps once and stopped recording
#### Version:
- Minecraft: 1.10.2
- Forge: 12.18.3.285
- Refined Storage: 1.2.24
Does this issue occur on a server? yes
| True | Extreme server lag with updating grid with new version of refinedstorage - Hello. I just upgraded refinedstorage on my server from version 1.2.16 to 1.2.24. However, in this version of refinedstorage, whenever I open any item grid: The grid opens instantly, but it does not show up to date items, The server freezes for 2-10 SECONDS, then the grid shows the items. There was no noticable lag in the older version of refinedstorage.
I was using ftb hermitpack with a couple of mods upgraded/added. (For a full list of changed mods, please refer to [this pastebin](http://pastebin.com/CvSQ2qqy) )
I have a midtier storage system with:
~60 crafters for crafting with recipies (I have to admit that many of them have ore dictionary, I was not aware of the performance consequences when I made them)
~20 crafters for machines
~20 external storage deep storage units
-3 disk drives
-The system is split over 2 remote locations in the same dimensions
The lag does not occur with fluid grids.
**I want to stress that I did not experience any lag in version 1.2.16 when opening the grid.**
I took some profiles of the lag:
[https://transfer.sh/EmO8n/refinedstorage.nps](https://transfer.sh/EmO8n/refinedstorage.nps)
I started recording, opened the crafting grid once, used /forge tps a couple times to see if the lag was gone, stopped recording
[https://transfer.sh/ZWqzQ/refinedstorage.nps](https://transfer.sh/ZWqzQ/refinedstorage.nps)
I started profiling, opened the crafting grid twice, used /forge tps once and stopped recording
#### Version:
- Minecraft: 1.10.2
- Forge: 12.18.3.285
- Refined Storage: 1.2.24
Does this issue occur on a server? yes
| non_priority | extreme server lag with updating grid with new version of refinedstorage hello i just upgraded refinedstorage on my server from version to however in this version of refinedstorage whenever i open any item grid the grid opens instantly but it does not show up to date items the server freezes for seconds then the grid shows the items there was no noticable lag in the older version of refinedstorage i was using ftb hermitpack with a couple of mods upgraded added for a full list of changed mods please refer to i have a midtier storage system with crafters for crafting with recipies i have to admit that many of them have ore dictionary i was not aware of the performance consequences when i made them crafters for machines external storage deep storage units disk drives the system is split over remote locations in the same dimensions the lag does not occur with fluid grids i want to stress that i did not experience any lag in version when opening the grid i took some profiles of the lag i started recording opened the crafting grid once used forge tps a couple times to see if the lag was gone stopped recording i started profiling opened the crafting grid twice used forge tps once and stopped recording version minecraft forge refined storage does this issue occur on a server yes | 0 |
425,049 | 29,189,827,191 | IssuesEvent | 2023-05-19 18:49:25 | aws/karpenter | https://api.github.com/repos/aws/karpenter | opened | docs: Add `consistency check failed with...` debugging info to Troubleshooting docs | documentation | ### Tell us about your request
Currently, users can see some eventing messages attached to their nodes due to consistency errors with Karpenter. These errors will appear like
```console
ERROR controller.consistency check failed, expected resource...
```
We should have information in our [Troubleshooting Guide](https://karpenter.sh/docs/troubleshooting) that covers what to do if you see errors like this and if there is any action to take from the user's side or if the error is benign. Currently, we don't have any docs on this so these errors can be misleading to some users who don't know what they mean.
### Tell us about the problem you're trying to solve. What are you trying to do, and why is it hard?
N/A
### Are you currently working around this issue?
N/A
### Additional Context
_No response_
### Attachments
_No response_
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
| 1.0 | docs: Add `consistency check failed with...` debugging info to Troubleshooting docs - ### Tell us about your request
Currently, users can see some eventing messages attached to their nodes due to consistency errors with Karpenter. These errors will appear like
```console
ERROR controller.consistency check failed, expected resource...
```
We should have information in our [Troubleshooting Guide](https://karpenter.sh/docs/troubleshooting) that covers what to do if you see errors like this and if there is any action to take from the user's side or if the error is benign. Currently, we don't have any docs on this so these errors can be misleading to some users who don't know what they mean.
### Tell us about the problem you're trying to solve. What are you trying to do, and why is it hard?
N/A
### Are you currently working around this issue?
N/A
### Additional Context
_No response_
### Attachments
_No response_
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
| non_priority | docs add consistency check failed with debugging info to troubleshooting docs tell us about your request currently users can see some eventing messages attached to their nodes due to consistency errors with karpenter these errors will appear like console error controller consistency check failed expected resource we should have information in our that covers what to do if you see errors like this and if there is any action to take from the user s side or if the error is benign currently we don t have any docs on this so these errors can be misleading to some users who don t know what they mean tell us about the problem you re trying to solve what are you trying to do and why is it hard n a are you currently working around this issue n a additional context no response attachments no response community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment | 0 |
10,820 | 9,106,325,516 | IssuesEvent | 2019-02-20 23:26:55 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | 3.0 SDK Documentation and .Net Core | app-service-web/svc assigned-to-author doc-bug triaged | Are these docs going to be updated for the 3.0 version of the SDK sometime soon. I'd love to use this to create WebJobs on .Net Core but I can't find any official docs on the Microsoft site for creating a .Net Core WebJob.
The closest thing to seeing anything new I've found is this sample https://github.com/Azure/azure-webjobs-sdk/blob/00686a5ae3b31ca1c70b477c1ca828e4aa754340/sample/SampleHost/Program.cs from a comment on this issue https://github.com/Azure/azure-webjobs-sdk/issues/1870#issuecomment-417888043
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a4500b14-64b4-4d1b-aae3-f0fd4dde85f2
* Version Independent ID: 8fb75741-37b9-b45f-c0ca-9948ddaf1c06
* Content: [How to use the WebJobs SDK - Azure](https://docs.microsoft.com/en-us/azure/app-service/webjobs-sdk-how-to#nextsteps)
* Content Source: [articles/app-service/webjobs-sdk-how-to.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/webjobs-sdk-how-to.md)
* Service: **app-service-web**
* GitHub Login: @ggailey777
* Microsoft Alias: **glenga** | 1.0 | 3.0 SDK Documentation and .Net Core - Are these docs going to be updated for the 3.0 version of the SDK sometime soon. I'd love to use this to create WebJobs on .Net Core but I can't find any official docs on the Microsoft site for creating a .Net Core WebJob.
The closest thing to seeing anything new I've found is this sample https://github.com/Azure/azure-webjobs-sdk/blob/00686a5ae3b31ca1c70b477c1ca828e4aa754340/sample/SampleHost/Program.cs from a comment on this issue https://github.com/Azure/azure-webjobs-sdk/issues/1870#issuecomment-417888043
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a4500b14-64b4-4d1b-aae3-f0fd4dde85f2
* Version Independent ID: 8fb75741-37b9-b45f-c0ca-9948ddaf1c06
* Content: [How to use the WebJobs SDK - Azure](https://docs.microsoft.com/en-us/azure/app-service/webjobs-sdk-how-to#nextsteps)
* Content Source: [articles/app-service/webjobs-sdk-how-to.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/webjobs-sdk-how-to.md)
* Service: **app-service-web**
* GitHub Login: @ggailey777
* Microsoft Alias: **glenga** | non_priority | sdk documentation and net core are these docs going to be updated for the version of the sdk sometime soon i d love to use this to create webjobs on net core but i can t find any official docs on the microsoft site for creating a net core webjob the closest thing to seeing anything new i ve found is this sample from a comment on this issue document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service web github login microsoft alias glenga | 0 |
32,392 | 7,531,106,197 | IssuesEvent | 2018-04-15 00:42:06 | dahall/TaskScheduler | https://api.github.com/repos/dahall/TaskScheduler | closed | Llaunching an un-elevated application from an elevated process | codeplex-disc | I have a WPF client application that checks against its server to ensure it is up to date. If it's not, an elevated process launches to perform the update and the client application closes. The updater re-launches the WPF client application after the update is complete, but it is launched with elevated privileges. I am trying to use the TaskService to get around that by having it launch the WPF client application with the leastpriviledges after the update is complete. I am seeing that it launches as a background task and is invisible to the user. I have checked Hidden = false on the task definition but it is still hidden. I have tried launching cmd.exe instead with the WPF client application as the arguments with the same result.
var arguments = string.Format("/C \"{0}\"", Executable);
using (var ts = new TaskService())
{
var taskName = "RelaunchApp";
var td = ts.NewTask();
td.RegistrationInfo.Description = string.Format("Relaunch application {0}", Executable);
td.Actions.Add(new ExecAction("cmd.exe", arguments));
td.Settings.Priority = ProcessPriorityClass.Normal;
td.Settings.Hidden = false;
td.Principal.RunLevel = TaskRunLevel.LUA;
td.Principal.LogonType = TaskLogonType.InteractiveToken;
td.Settings.AllowDemandStart = true;
td.Settings.DisallowStartIfOnBatteries = false;
td.Settings.StopIfGoingOnBatteries = false;
var ret = ts.RootFolder.RegisterTaskDefinition(taskName, td);
var clientTask = ts.FindTask(taskName);
if (null != clientTask)
{
if (clientTask.Enabled)
{
clientTask.Run(arguments);
}
}
else
{
TraceLine("Task not found.");
}
td.Dispose();
ts.RootFolder.DeleteTask(taskName);
}
I have also tried launching the WPF client app directly:
using (var ts = new TaskService())
{
var taskName = "RelaunchApp";
var td = ts.NewTask();
td.RegistrationInfo.Description = string.Format("Relaunch application {0}", Executable);
td.Actions.Add(new ExecAction(Executable));
td.Settings.Priority = ProcessPriorityClass.Normal;
td.Settings.Hidden = false;
td.Principal.RunLevel = TaskRunLevel.LUA;
td.Settings.AllowDemandStart = true;
td.Settings.DisallowStartIfOnBatteries = false;
td.Settings.StopIfGoingOnBatteries = false;
var ret = ts.RootFolder.RegisterTaskDefinition(taskName, td);
var clientTask = ts.FindTask(taskName);
if (null != clientTask)
{
if (clientTask.Enabled)
{
clientTask.Run();
}
}
else
{
TraceLine("Task not found.");
}
td.Dispose();
ts.RootFolder.DeleteTask(taskName);
}
What do I need to do to make the application visible?
Thank you!
Originally posted: 2016-07-13T11:48:16 | 1.0 | Llaunching an un-elevated application from an elevated process - I have a WPF client application that checks against its server to ensure it is up to date. If it's not, an elevated process launches to perform the update and the client application closes. The updater re-launches the WPF client application after the update is complete, but it is launched with elevated privileges. I am trying to use the TaskService to get around that by having it launch the WPF client application with the leastpriviledges after the update is complete. I am seeing that it launches as a background task and is invisible to the user. I have checked Hidden = false on the task definition but it is still hidden. I have tried launching cmd.exe instead with the WPF client application as the arguments with the same result.
var arguments = string.Format("/C \"{0}\"", Executable);
using (var ts = new TaskService())
{
var taskName = "RelaunchApp";
var td = ts.NewTask();
td.RegistrationInfo.Description = string.Format("Relaunch application {0}", Executable);
td.Actions.Add(new ExecAction("cmd.exe", arguments));
td.Settings.Priority = ProcessPriorityClass.Normal;
td.Settings.Hidden = false;
td.Principal.RunLevel = TaskRunLevel.LUA;
td.Principal.LogonType = TaskLogonType.InteractiveToken;
td.Settings.AllowDemandStart = true;
td.Settings.DisallowStartIfOnBatteries = false;
td.Settings.StopIfGoingOnBatteries = false;
var ret = ts.RootFolder.RegisterTaskDefinition(taskName, td);
var clientTask = ts.FindTask(taskName);
if (null != clientTask)
{
if (clientTask.Enabled)
{
clientTask.Run(arguments);
}
}
else
{
TraceLine("Task not found.");
}
td.Dispose();
ts.RootFolder.DeleteTask(taskName);
}
I have also tried launching the WPF client app directly:
using (var ts = new TaskService())
{
var taskName = "RelaunchApp";
var td = ts.NewTask();
td.RegistrationInfo.Description = string.Format("Relaunch application {0}", Executable);
td.Actions.Add(new ExecAction(Executable));
td.Settings.Priority = ProcessPriorityClass.Normal;
td.Settings.Hidden = false;
td.Principal.RunLevel = TaskRunLevel.LUA;
td.Settings.AllowDemandStart = true;
td.Settings.DisallowStartIfOnBatteries = false;
td.Settings.StopIfGoingOnBatteries = false;
var ret = ts.RootFolder.RegisterTaskDefinition(taskName, td);
var clientTask = ts.FindTask(taskName);
if (null != clientTask)
{
if (clientTask.Enabled)
{
clientTask.Run();
}
}
else
{
TraceLine("Task not found.");
}
td.Dispose();
ts.RootFolder.DeleteTask(taskName);
}
What do I need to do to make the application visible?
Thank you!
Originally posted: 2016-07-13T11:48:16 | non_priority | llaunching an un elevated application from an elevated process i have a wpf client application that checks against its server to ensure it is up to date if it s not an elevated process launches to perform the update and the client application closes the updater re launches the wpf client application after the update is complete but it is launched with elevated privileges i am trying to use the taskservice to get around that by having it launch the wpf client application with the leastpriviledges after the update is complete i am seeing that it launches as a background task and is invisible to the user i have checked hidden false on the task definition but it is still hidden i have tried launching cmd exe instead with the wpf client application as the arguments with the same result var arguments string format quot c quot quot quot executable using var ts new taskservice var taskname quot relaunchapp quot var td ts newtask td registrationinfo description string format quot relaunch application quot executable td actions add new execaction quot cmd exe quot arguments td settings priority processpriorityclass normal td settings hidden false td principal runlevel taskrunlevel lua td principal logontype tasklogontype interactivetoken td settings allowdemandstart true td settings disallowstartifonbatteries false td settings stopifgoingonbatteries false var ret ts rootfolder registertaskdefinition taskname td var clienttask ts findtask taskname if null clienttask if clienttask enabled clienttask run arguments else traceline quot task not found quot td dispose ts rootfolder deletetask taskname i have also tried launching the wpf client app directly using var ts new taskservice var taskname quot relaunchapp quot var td ts newtask td registrationinfo description string format quot relaunch application quot executable td actions add new execaction executable td settings priority processpriorityclass normal td settings hidden false td principal runlevel taskrunlevel lua td settings allowdemandstart true td settings disallowstartifonbatteries false td settings stopifgoingonbatteries false var ret ts rootfolder registertaskdefinition taskname td var clienttask ts findtask taskname if null clienttask if clienttask enabled clienttask run else traceline quot task not found quot td dispose ts rootfolder deletetask taskname what do i need to do to make the application visible thank you originally posted | 0 |
26,083 | 26,393,498,302 | IssuesEvent | 2023-01-12 17:23:43 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Ctrl+S does not save resources in *[unsaved] scenes | enhancement topic:editor usability |
**Godot version:**
4.0.dev c9413b513fd2fb95e9b71d886e01a04e379bb969
**OS/device including version:**
Windows 10.0.19041.746
Nvidia RTX 2080 Ti driver 452.06
**Issue description:**
Ctrl+S saves resources but only if the current scene can be saved, so either [empty] or a valid scene. (not [unsaved], and not opened .glb files or such things)
Ctrl+Alt+S (Save All Scenes) always saves resources but does not notify the user of this fact.
Similar to issue 9069, but this problem deals with *all resources* (such as Environment), not just scripts.
**Steps to reproduce:**
1. Create a new project or open existing
2. Double click default_env.tres
3. Expand Sky. Then, in the Sky fleld, click Sky to expand the sky object.
4. In the Sky Material box, create a new PhysicalSky
5. Press Ctrl-S
6. Quit to Project List and reopen.
7. The default_sky.tres was not saved.
**Minimal reproduction project:**
Happens in a new project | True | Ctrl+S does not save resources in *[unsaved] scenes -
**Godot version:**
4.0.dev c9413b513fd2fb95e9b71d886e01a04e379bb969
**OS/device including version:**
Windows 10.0.19041.746
Nvidia RTX 2080 Ti driver 452.06
**Issue description:**
Ctrl+S saves resources but only if the current scene can be saved, so either [empty] or a valid scene. (not [unsaved], and not opened .glb files or such things)
Ctrl+Alt+S (Save All Scenes) always saves resources but does not notify the user of this fact.
Similar to issue 9069, but this problem deals with *all resources* (such as Environment), not just scripts.
**Steps to reproduce:**
1. Create a new project or open existing
2. Double click default_env.tres
3. Expand Sky. Then, in the Sky fleld, click Sky to expand the sky object.
4. In the Sky Material box, create a new PhysicalSky
5. Press Ctrl-S
6. Quit to Project List and reopen.
7. The default_sky.tres was not saved.
**Minimal reproduction project:**
Happens in a new project | non_priority | ctrl s does not save resources in scenes godot version dev os device including version windows nvidia rtx ti driver issue description ctrl s saves resources but only if the current scene can be saved so either or a valid scene not and not opened glb files or such things ctrl alt s save all scenes always saves resources but does not notify the user of this fact similar to issue but this problem deals with all resources such as environment not just scripts steps to reproduce create a new project or open existing double click default env tres expand sky then in the sky fleld click sky to expand the sky object in the sky material box create a new physicalsky press ctrl s quit to project list and reopen the default sky tres was not saved minimal reproduction project happens in a new project | 0 |
74,872 | 20,386,732,969 | IssuesEvent | 2022-02-22 07:53:18 | buildingSMART/IFC4.3.x-development | https://api.github.com/repos/buildingSMART/IFC4.3.x-development | closed | Purge all references to standardcase in the docs | allocated-core decided allocated-building | Originally brought up [here](https://forums.buildingsmart.org/t/how-to-contribute-to-ifc-documentation/3161?u=moult)
I grepped `docs/{schemas,templates}` for "standard" and found lots of mentions of standard case which needs vetting. I'm happy to go through and do a purge if everyone is OK for it. Naturally the purge will be documented in a commit for review in case I make an error in judgement :)
Let me know if OK to go ahead! | 1.0 | Purge all references to standardcase in the docs - Originally brought up [here](https://forums.buildingsmart.org/t/how-to-contribute-to-ifc-documentation/3161?u=moult)
I grepped `docs/{schemas,templates}` for "standard" and found lots of mentions of standard case which needs vetting. I'm happy to go through and do a purge if everyone is OK for it. Naturally the purge will be documented in a commit for review in case I make an error in judgement :)
Let me know if OK to go ahead! | non_priority | purge all references to standardcase in the docs originally brought up i grepped docs schemas templates for standard and found lots of mentions of standard case which needs vetting i m happy to go through and do a purge if everyone is ok for it naturally the purge will be documented in a commit for review in case i make an error in judgement let me know if ok to go ahead | 0 |
110,863 | 11,711,721,568 | IssuesEvent | 2020-03-09 06:16:12 | SE701ProjectGroup4/Flatmate-Management-System | https://api.github.com/repos/SE701ProjectGroup4/Flatmate-Management-System | opened | Create Pull Request Template | documentation | As a Developer, I want a standardized method to check the branch is in the correct state before creating a pull request
Acceptance Criteria:
- The format is documented on Github wiki | 1.0 | Create Pull Request Template - As a Developer, I want a standardized method to check the branch is in the correct state before creating a pull request
Acceptance Criteria:
- The format is documented on Github wiki | non_priority | create pull request template as a developer i want a standardized method to check the branch is in the correct state before creating a pull request acceptance criteria the format is documented on github wiki | 0 |
91,082 | 11,462,152,664 | IssuesEvent | 2020-02-07 13:34:57 | sugarlabs/musicblocks | https://api.github.com/repos/sugarlabs/musicblocks | closed | Redesign Widgets | Component-Widget Issue-Design | How do we make the widgets more intuitive? Work better within the other aspects of MB's UI?
This discussion refers to some challenges observed with users in the classroom: https://github.com/sugarlabs/musicblocks/issues/201#issuecomment-437744528
(some have been solved)
Do we want Full screen?
https://github.com/sugarlabs/musicblocks/issues/1379
And if not, how do we make it more suggestive to the user where the play/stop functions are.
First step is to propose a design. Visual mockups would be helpful. | 1.0 | Redesign Widgets - How do we make the widgets more intuitive? Work better within the other aspects of MB's UI?
This discussion refers to some challenges observed with users in the classroom: https://github.com/sugarlabs/musicblocks/issues/201#issuecomment-437744528
(some have been solved)
Do we want Full screen?
https://github.com/sugarlabs/musicblocks/issues/1379
And if not, how do we make it more suggestive to the user where the play/stop functions are.
First step is to propose a design. Visual mockups would be helpful. | non_priority | redesign widgets how do we make the widgets more intuitive work better within the other aspects of mb s ui this discussion refers to some challenges observed with users in the classroom some have been solved do we want full screen and if not how do we make it more suggestive to the user where the play stop functions are first step is to propose a design visual mockups would be helpful | 0 |
21,341 | 14,530,472,296 | IssuesEvent | 2020-12-14 19:18:55 | microsoft/react-native-windows | https://api.github.com/repos/microsoft/react-native-windows | closed | CI Loop: tree dump comparison file not being uploaded as artifacts | Area: Test Infrastructure bug | @chiaramooney ran into this issue while working on:
https://github.com/microsoft/react-native-windows/pull/6609
The problem we're seeing is that when one of the E2E tests fails a master treedump comparison, the comparison file is not being uploaded as a test artifact. This makes it impossible to see what actually failed by looking at the logs.
I created a little PR to repro this issue here:
https://github.com/microsoft/react-native-windows/pull/6683
Go to the CI loop ADO page and dig into the E2E test failure. Notice you cannot locate the test comparison file, which is supposed to be here:

Chiara and I debugged together and can see that when we run the test locally, the expected comparison file is generated on the local box. This makes me think the issue is somewhere in the artifact publishing process.
| 1.0 | CI Loop: tree dump comparison file not being uploaded as artifacts - @chiaramooney ran into this issue while working on:
https://github.com/microsoft/react-native-windows/pull/6609
The problem we're seeing is that when one of the E2E tests fails a master treedump comparison, the comparison file is not being uploaded as a test artifact. This makes it impossible to see what actually failed by looking at the logs.
I created a little PR to repro this issue here:
https://github.com/microsoft/react-native-windows/pull/6683
Go to the CI loop ADO page and dig into the E2E test failure. Notice you cannot locate the test comparison file, which is supposed to be here:

Chiara and I debugged together and can see that when we run the test locally, the expected comparison file is generated on the local box. This makes me think the issue is somewhere in the artifact publishing process.
| non_priority | ci loop tree dump comparison file not being uploaded as artifacts chiaramooney ran into this issue while working on the problem we re seeing is that when one of the tests fails a master treedump comparison the comparison file is not being uploaded as a test artifact this makes it impossible to see what actually failed by looking at the logs i created a little pr to repro this issue here go to the ci loop ado page and dig into the test failure notice you cannot locate the test comparison file which is supposed to be here chiara and i debugged together and can see that when we run the test locally the expected comparison file is generated on the local box this makes me think the issue is somewhere in the artifact publishing process | 0 |
211 | 2,552,182,872 | IssuesEvent | 2015-02-02 15:40:04 | opattison/olivermakes | https://api.github.com/repos/opattison/olivermakes | opened | Use font language subsetting in Typekit to save on file size | low-hanging fruit performance | ## Advantages
- ~20% savings (for three weights of Ratio); ~28% savings (for Ratio and Source Code together)
- Maintains OpenType features
- Can be reconfigured later if more language support is needed
## Disadvantages
None.
OpenType features are now a checkbox in the kit configuration panel. | True | Use font language subsetting in Typekit to save on file size - ## Advantages
- ~20% savings (for three weights of Ratio); ~28% savings (for Ratio and Source Code together)
- Maintains OpenType features
- Can be reconfigured later if more language support is needed
## Disadvantages
None.
OpenType features are now a checkbox in the kit configuration panel. | non_priority | use font language subsetting in typekit to save on file size advantages savings for three weights of ratio savings for ratio and source code together maintains opentype features can be reconfigured later if more language support is needed disadvantages none opentype features are now a checkbox in the kit configuration panel | 0 |
405,442 | 27,519,226,724 | IssuesEvent | 2023-03-06 14:03:34 | jphayek/Git-tp-finale | https://api.github.com/repos/jphayek/Git-tp-finale | closed | Github plus | documentation enhancement | Github Plus is a chrome extension aimed at giving developers more tools when working on the github website
| 1.0 | Github plus - Github Plus is a chrome extension aimed at giving developers more tools when working on the github website
| non_priority | github plus github plus is a chrome extension aimed at giving developers more tools when working on the github website | 0 |
39,705 | 16,068,128,579 | IssuesEvent | 2021-04-23 23:32:26 | provenance-io/explorer-service | https://api.github.com/repos/provenance-io/explorer-service | closed | Add pagination to API calls | service functionality | <!-- --------------------------------------------------------
Thank you for opening an issue.
Before submitting this request please review this template.
-------------------------------------------------------- -->
## Summary
Add pagination to API calls
## Problem Definition
<!-- Why do we need this feature?
What problems may be addressed by introducing this feature?
What benefits does the Provenance Network Explorer stand to gain by including this feature?
Does this require support from Provenance Blockchain or Cosmos SDK? Consider submitting a request
upstream as well.
Are there any disadvantages of including this feature? -->
These calls are used to populate the frontend UI. To make the UI better, pagination needs to be a thing, and this means APIs need to support pagination as well.
## Proposal
There is existing pagination in place for recent transactions and validators. This should be applied to any API that returns a list of objects.
____
#### For Admin Use
- [x] Not duplicate issue
- [x] Appropriate labels applied
- [x] Appropriate contributors tagged
- [x] Contributor assigned/self-assigned
| 1.0 | Add pagination to API calls - <!-- --------------------------------------------------------
Thank you for opening an issue.
Before submitting this request please review this template.
-------------------------------------------------------- -->
## Summary
Add pagination to API calls
## Problem Definition
<!-- Why do we need this feature?
What problems may be addressed by introducing this feature?
What benefits does the Provenance Network Explorer stand to gain by including this feature?
Does this require support from Provenance Blockchain or Cosmos SDK? Consider submitting a request
upstream as well.
Are there any disadvantages of including this feature? -->
These calls are used to populate the frontend UI. To make the UI better, pagination needs to be a thing, and this means APIs need to support pagination as well.
## Proposal
There is existing pagination in place for recent transactions and validators. This should be applied to any API that returns a list of objects.
____
#### For Admin Use
- [x] Not duplicate issue
- [x] Appropriate labels applied
- [x] Appropriate contributors tagged
- [x] Contributor assigned/self-assigned
| non_priority | add pagination to api calls thank you for opening an issue before submitting this request please review this template summary add pagination to api calls problem definition why do we need this feature what problems may be addressed by introducing this feature what benefits does the provenance network explorer stand to gain by including this feature does this require support from provenance blockchain or cosmos sdk consider submitting a request upstream as well are there any disadvantages of including this feature these calls are used to populate the frontend ui to make the ui better pagination needs to be a thing and this means apis need to support pagination as well proposal there is existing pagination in place for recent transactions and validators this should be applied to any api that returns a list of objects for admin use not duplicate issue appropriate labels applied appropriate contributors tagged contributor assigned self assigned | 0 |
151,775 | 19,665,342,464 | IssuesEvent | 2022-01-10 21:46:20 | tom9carthron1/infinite-wish-board | https://api.github.com/repos/tom9carthron1/infinite-wish-board | closed | WS-2019-0318 (High) detected in handlebars-4.4.1.tgz - autoclosed | security vulnerability | ## WS-2019-0318 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.4.1.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.1.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.1.tgz</a></p>
<p>Path to dependency file: infinite-wish-board/ui/package.json</p>
<p>Path to vulnerable library: infinite-wish-board/ui/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- jest-24.7.1.tgz
- jest-cli-24.9.0.tgz
- core-24.9.0.tgz
- reporters-24.9.0.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.4.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tom9carthron1/infinite-wish-board/commit/5c98c8171c2f014c3587805f1090b7d16419c286">5c98c8171c2f014c3587805f1090b7d16419c286</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.
<p>Publish Date: 2019-10-20
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.4.1","packageFilePaths":["/ui/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:3.0.1;jest:24.7.1;jest-cli:24.9.0;@jest/core:24.9.0;@jest/reporters:24.9.0;istanbul-reports:2.2.6;handlebars:4.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.4.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0318","vulnerabilityDetails":"In \"showdownjs/showdown\", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2019-0318 (High) detected in handlebars-4.4.1.tgz - autoclosed - ## WS-2019-0318 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.4.1.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.1.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.1.tgz</a></p>
<p>Path to dependency file: infinite-wish-board/ui/package.json</p>
<p>Path to vulnerable library: infinite-wish-board/ui/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- jest-24.7.1.tgz
- jest-cli-24.9.0.tgz
- core-24.9.0.tgz
- reporters-24.9.0.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.4.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tom9carthron1/infinite-wish-board/commit/5c98c8171c2f014c3587805f1090b7d16419c286">5c98c8171c2f014c3587805f1090b7d16419c286</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.
<p>Publish Date: 2019-10-20
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.4.1","packageFilePaths":["/ui/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:3.0.1;jest:24.7.1;jest-cli:24.9.0;@jest/core:24.9.0;@jest/reporters:24.9.0;istanbul-reports:2.2.6;handlebars:4.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.4.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0318","vulnerabilityDetails":"In \"showdownjs/showdown\", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | ws high detected in handlebars tgz autoclosed ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file infinite wish board ui package json path to vulnerable library infinite wish board ui node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz core tgz reporters tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details in showdownjs showdown versions prior to are vulnerable against regular expression denial of service redos once receiving specially crafted templates publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree react scripts jest jest cli jest core jest reporters istanbul reports handlebars isminimumfixversionavailable true minimumfixversion handlebars basebranches vulnerabilityidentifier ws vulnerabilitydetails in showdownjs showdown versions prior to are vulnerable against regular expression denial of service redos once receiving specially crafted templates vulnerabilityurl | 0 |
223,294 | 24,711,752,030 | IssuesEvent | 2022-10-20 01:45:11 | turkdevops/prettier | https://api.github.com/repos/turkdevops/prettier | closed | WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed | security vulnerability | ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/rollup-plugin-node-globals/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- rollup-plugin-node-globals-1.4.0.tgz (Root Library)
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/prettier/commit/4bcfa0c3b57b624037b4ed29391d87f76e57a7f3">4bcfa0c3b57b624037b4ed29391d87f76e57a7f3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/rollup-plugin-node-globals/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- rollup-plugin-node-globals-1.4.0.tgz (Root Library)
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/prettier/commit/4bcfa0c3b57b624037b4ed29391d87f76e57a7f3">4bcfa0c3b57b624037b4ed29391d87f76e57a7f3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws high detected in acorn tgz autoclosed ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file package json path to vulnerable library node modules rollup plugin node globals node modules acorn package json dependency hierarchy rollup plugin node globals tgz root library x acorn tgz vulnerable library found in head commit a href found in base branch master vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
45,788 | 11,733,429,444 | IssuesEvent | 2020-03-11 07:04:46 | mapbox/mapbox-navigation-ios | https://api.github.com/repos/mapbox/mapbox-navigation-ios | closed | jazzy table of contents specifies nonexistent symbols | - build documentation topic: documentation | Running document.sh produces the following warnings due to stale entries in the [jazzy table of contents](https://github.com/mapbox/mapbox-navigation-ios/blob/2671f08a7acf172d598436be96f4228d232a3e6f/docs/jazzy.yml#L14-L89):
```
WARNING: No documented top-level declarations match name "RouteControllerDelegate" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerNotificationUserInfoKey" specified in categories file
WARNING: No documented top-level declarations match name "StyleKitArrows" specified in categories file
WARNING: No documented top-level declarations match name "StylableButton" specified in categories file
WARNING: No documented top-level declarations match name "StylableLabel" specified in categories file
WARNING: No documented top-level declarations match name "RouteStepFormatter" specified in categories file
WARNING: No documented top-level declarations match name "RouteTableViewController" specified in categories file
WARNING: No documented top-level declarations match name "NavigationPulleyViewController" specified in categories file
WARNING: No documented top-level declarations match name "Button" specified in categories file
WARNING: No documented top-level declarations match name "CellSubtitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "CellTitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "DashedLineView" specified in categories file
WARNING: No documented top-level declarations match name "HighlightedButton" specified in categories file
WARNING: No documented top-level declarations match name "LineView" specified in categories file
WARNING: No documented top-level declarations match name "ProgressBar" specified in categories file
WARNING: No documented top-level declarations match name "SeparatorView" specified in categories file
WARNING: No documented top-level declarations match name "SubtitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "TitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "TurnArrowView" specified in categories file
WARNING: No documented top-level declarations match name "WayNameLabel" specified in categories file
WARNING: No documented top-level declarations match name "MaxSecondsSpentTravelingAwayFromStartOfRoute" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMaxManipulatedCourseAngle" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForHighAlertCycling" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForHighAlertDriving" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForHighAlertWalking" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForMediumAlertCycling" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForMediumAlertDriving" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForMediumAlertWalking" specified in categories file
```
We should remove any entries that are no longer present in the SDK and rename any entries that were renamed in PRs such as #2230.
/cc @mapbox/navigation-ios | 1.0 | jazzy table of contents specifies nonexistent symbols - Running document.sh produces the following warnings due to stale entries in the [jazzy table of contents](https://github.com/mapbox/mapbox-navigation-ios/blob/2671f08a7acf172d598436be96f4228d232a3e6f/docs/jazzy.yml#L14-L89):
```
WARNING: No documented top-level declarations match name "RouteControllerDelegate" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerNotificationUserInfoKey" specified in categories file
WARNING: No documented top-level declarations match name "StyleKitArrows" specified in categories file
WARNING: No documented top-level declarations match name "StylableButton" specified in categories file
WARNING: No documented top-level declarations match name "StylableLabel" specified in categories file
WARNING: No documented top-level declarations match name "RouteStepFormatter" specified in categories file
WARNING: No documented top-level declarations match name "RouteTableViewController" specified in categories file
WARNING: No documented top-level declarations match name "NavigationPulleyViewController" specified in categories file
WARNING: No documented top-level declarations match name "Button" specified in categories file
WARNING: No documented top-level declarations match name "CellSubtitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "CellTitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "DashedLineView" specified in categories file
WARNING: No documented top-level declarations match name "HighlightedButton" specified in categories file
WARNING: No documented top-level declarations match name "LineView" specified in categories file
WARNING: No documented top-level declarations match name "ProgressBar" specified in categories file
WARNING: No documented top-level declarations match name "SeparatorView" specified in categories file
WARNING: No documented top-level declarations match name "SubtitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "TitleLabel" specified in categories file
WARNING: No documented top-level declarations match name "TurnArrowView" specified in categories file
WARNING: No documented top-level declarations match name "WayNameLabel" specified in categories file
WARNING: No documented top-level declarations match name "MaxSecondsSpentTravelingAwayFromStartOfRoute" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMaxManipulatedCourseAngle" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForHighAlertCycling" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForHighAlertDriving" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForHighAlertWalking" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForMediumAlertCycling" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForMediumAlertDriving" specified in categories file
WARNING: No documented top-level declarations match name "RouteControllerMinimumDistanceForMediumAlertWalking" specified in categories file
```
We should remove any entries that are no longer present in the SDK and rename any entries that were renamed in PRs such as #2230.
/cc @mapbox/navigation-ios | non_priority | jazzy table of contents specifies nonexistent symbols running document sh produces the following warnings due to stale entries in the warning no documented top level declarations match name routecontrollerdelegate specified in categories file warning no documented top level declarations match name routecontrollernotificationuserinfokey specified in categories file warning no documented top level declarations match name stylekitarrows specified in categories file warning no documented top level declarations match name stylablebutton specified in categories file warning no documented top level declarations match name stylablelabel specified in categories file warning no documented top level declarations match name routestepformatter specified in categories file warning no documented top level declarations match name routetableviewcontroller specified in categories file warning no documented top level declarations match name navigationpulleyviewcontroller specified in categories file warning no documented top level declarations match name button specified in categories file warning no documented top level declarations match name cellsubtitlelabel specified in categories file warning no documented top level declarations match name celltitlelabel specified in categories file warning no documented top level declarations match name dashedlineview specified in categories file warning no documented top level declarations match name highlightedbutton specified in categories file warning no documented top level declarations match name lineview specified in categories file warning no documented top level declarations match name progressbar specified in categories file warning no documented top level declarations match name separatorview specified in categories file warning no documented top level declarations match name subtitlelabel specified in categories file warning no documented top level declarations match name titlelabel specified in categories file warning no documented top level declarations match name turnarrowview specified in categories file warning no documented top level declarations match name waynamelabel specified in categories file warning no documented top level declarations match name maxsecondsspenttravelingawayfromstartofroute specified in categories file warning no documented top level declarations match name routecontrollermaxmanipulatedcourseangle specified in categories file warning no documented top level declarations match name routecontrollerminimumdistanceforhighalertcycling specified in categories file warning no documented top level declarations match name routecontrollerminimumdistanceforhighalertdriving specified in categories file warning no documented top level declarations match name routecontrollerminimumdistanceforhighalertwalking specified in categories file warning no documented top level declarations match name routecontrollerminimumdistanceformediumalertcycling specified in categories file warning no documented top level declarations match name routecontrollerminimumdistanceformediumalertdriving specified in categories file warning no documented top level declarations match name routecontrollerminimumdistanceformediumalertwalking specified in categories file we should remove any entries that are no longer present in the sdk and rename any entries that were renamed in prs such as cc mapbox navigation ios | 0 |
174,754 | 13,518,875,075 | IssuesEvent | 2020-09-15 00:21:49 | rancher/fleet | https://api.github.com/repos/rancher/fleet | closed | Failed to deploy apps into clusters in Rancher single-install setup | [zube]: To Test bug status/blocker | the clusters are provisioned by Rancher
```
> k -n clusters get clusters -o=jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.metadata.labels}{"\n"}{end}'
cluster-dc8270631e5b map[env:dev fleet.cattle.io/cluster:cluster-dc8270631e5b]
cluster-ddff1b5b592d map[env:test fleet.cattle.io/cluster:cluster-ddff1b5b592d]
cluster-fc9e80ee366b map[env:prod fleet.cattle.io/cluster:cluster-fc9e80ee366b]
```
use the following yaml to deploy a `GitRepo` CRD (source https://fleet.rancher.io/gitrepo-targets/)
```
kind: GitRepo
apiVersion: fleet.cattle.io/v1alpha1
metadata:
name: myrepo
namespace: clusters
spec:
repo: http://github.com/rancher/fleet-examples
bundleDirs:
- simple
# Targets are evaluated in order and the first one to match is used. If
# no targets match then the evaluated cluster will not be deployed to.
targets:
# The name of target. If not specified a default name of the format "target000"
# will be used
- name: prod
# A selector used to match clusters. The structure is the standard
# metav1.LabelSelector format. If clusterGroupSelector or clusterGroup is specified,
# clusterSelector will be used only to further refine the selection after
# clusterGroupSelector and clusterGroup is evaluated.
clusterSelector:
matchLabels:
env: prod
# A selector used to match cluster groups.
clusterGroupSelector:
matchLabels:
region: us-east
# A specific clusterGroup by name that will be selected
clusterGroup: group1
```
Results:
```
> k -n clusters get all
NAME READY STATUS RESTARTS AGE
pod/myrepo-c6af9-8lhfb 0/2 Error 0 2m24s
pod/myrepo-c6af9-cb7sk 0/2 Error 0 7m25s
pod/myrepo-c6af9-gx5jq 0/2 Error 0 7m43s
pod/myrepo-c6af9-ktlfw 0/2 Error 0 5m4s
pod/myrepo-c6af9-ltvd6 0/2 Error 0 7m5s
pod/myrepo-c6af9-n5h99 0/2 Error 0 6m25s
pod/myrepo-c6af9-wzf7m 0/2 Error 0 7m35s
NAME COMPLETIONS DURATION AGE
job.batch/myrepo-c6af9 0/1 7m44s 7m44s
```
log of the container `step-git-source`
`{"level":"error","ts":1599093912.062322,"caller":"git/git.go:41","msg":"Error running git [fetch --recurse-submodules=yes --depth=1 origin --update-head-ok --force warning:]: exit status 128\nwarning: redirecting to https://github.com/rancher/fleet-examples/\nfatal: couldn't find remote ref warning\n","stacktrace":"github.com/tektoncd/pipeline/pkg/git.run\n\tgithub.com/tektoncd/pipeline/pkg/git/git.go:41\ngithub.com/tektoncd/pipeline/pkg/git.Fetch\n\tgithub.com/tektoncd/pipeline/pkg/git/git.go:116\nmain.main\n\tgithub.com/tektoncd/pipeline/cmd/git-init/main.go:53\nruntime.main\n\truntime/proc.go:203"}
{"level":"fatal","ts":1599093912.0624912,"caller":"git-init/main.go:54","msg":"Error fetching git repository: failed to fetch [warning:]: exit status 128","stacktrace":"main.main\n\tgithub.com/tektoncd/pipeline/cmd/git-init/main.go:54\nruntime.main\n\truntime/proc.go:203"}`
<img width="1463" alt="Screen Shot 2020-09-02 at 5 48 40 PM" src="https://user-images.githubusercontent.com/6218999/92059086-1f303200-ed45-11ea-9ac0-6cdcae6a496d.png">
--------
Update:
It turns out this bug happens only in the Rancher single-install setup | 1.0 | Failed to deploy apps into clusters in Rancher single-install setup - the clusters are provisioned by Rancher
```
> k -n clusters get clusters -o=jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.metadata.labels}{"\n"}{end}'
cluster-dc8270631e5b map[env:dev fleet.cattle.io/cluster:cluster-dc8270631e5b]
cluster-ddff1b5b592d map[env:test fleet.cattle.io/cluster:cluster-ddff1b5b592d]
cluster-fc9e80ee366b map[env:prod fleet.cattle.io/cluster:cluster-fc9e80ee366b]
```
use the following yaml to deploy a `GitRepo` CRD (source https://fleet.rancher.io/gitrepo-targets/)
```
kind: GitRepo
apiVersion: fleet.cattle.io/v1alpha1
metadata:
name: myrepo
namespace: clusters
spec:
repo: http://github.com/rancher/fleet-examples
bundleDirs:
- simple
# Targets are evaluated in order and the first one to match is used. If
# no targets match then the evaluated cluster will not be deployed to.
targets:
# The name of target. If not specified a default name of the format "target000"
# will be used
- name: prod
# A selector used to match clusters. The structure is the standard
# metav1.LabelSelector format. If clusterGroupSelector or clusterGroup is specified,
# clusterSelector will be used only to further refine the selection after
# clusterGroupSelector and clusterGroup is evaluated.
clusterSelector:
matchLabels:
env: prod
# A selector used to match cluster groups.
clusterGroupSelector:
matchLabels:
region: us-east
# A specific clusterGroup by name that will be selected
clusterGroup: group1
```
Results:
```
> k -n clusters get all
NAME READY STATUS RESTARTS AGE
pod/myrepo-c6af9-8lhfb 0/2 Error 0 2m24s
pod/myrepo-c6af9-cb7sk 0/2 Error 0 7m25s
pod/myrepo-c6af9-gx5jq 0/2 Error 0 7m43s
pod/myrepo-c6af9-ktlfw 0/2 Error 0 5m4s
pod/myrepo-c6af9-ltvd6 0/2 Error 0 7m5s
pod/myrepo-c6af9-n5h99 0/2 Error 0 6m25s
pod/myrepo-c6af9-wzf7m 0/2 Error 0 7m35s
NAME COMPLETIONS DURATION AGE
job.batch/myrepo-c6af9 0/1 7m44s 7m44s
```
log of the container `step-git-source`
`{"level":"error","ts":1599093912.062322,"caller":"git/git.go:41","msg":"Error running git [fetch --recurse-submodules=yes --depth=1 origin --update-head-ok --force warning:]: exit status 128\nwarning: redirecting to https://github.com/rancher/fleet-examples/\nfatal: couldn't find remote ref warning\n","stacktrace":"github.com/tektoncd/pipeline/pkg/git.run\n\tgithub.com/tektoncd/pipeline/pkg/git/git.go:41\ngithub.com/tektoncd/pipeline/pkg/git.Fetch\n\tgithub.com/tektoncd/pipeline/pkg/git/git.go:116\nmain.main\n\tgithub.com/tektoncd/pipeline/cmd/git-init/main.go:53\nruntime.main\n\truntime/proc.go:203"}
{"level":"fatal","ts":1599093912.0624912,"caller":"git-init/main.go:54","msg":"Error fetching git repository: failed to fetch [warning:]: exit status 128","stacktrace":"main.main\n\tgithub.com/tektoncd/pipeline/cmd/git-init/main.go:54\nruntime.main\n\truntime/proc.go:203"}`
<img width="1463" alt="Screen Shot 2020-09-02 at 5 48 40 PM" src="https://user-images.githubusercontent.com/6218999/92059086-1f303200-ed45-11ea-9ac0-6cdcae6a496d.png">
--------
Update:
It turns out this bug happens only in the Rancher single-install setup | non_priority | failed to deploy apps into clusters in rancher single install setup the clusters are provisioned by rancher k n clusters get clusters o jsonpath range items metadata name t metadata labels n end cluster map cluster map cluster map use the following yaml to deploy a gitrepo crd source kind gitrepo apiversion fleet cattle io metadata name myrepo namespace clusters spec repo bundledirs simple targets are evaluated in order and the first one to match is used if no targets match then the evaluated cluster will not be deployed to targets the name of target if not specified a default name of the format will be used name prod a selector used to match clusters the structure is the standard labelselector format if clustergroupselector or clustergroup is specified clusterselector will be used only to further refine the selection after clustergroupselector and clustergroup is evaluated clusterselector matchlabels env prod a selector used to match cluster groups clustergroupselector matchlabels region us east a specific clustergroup by name that will be selected clustergroup results k n clusters get all name ready status restarts age pod myrepo error pod myrepo error pod myrepo error pod myrepo ktlfw error pod myrepo error pod myrepo error pod myrepo error name completions duration age job batch myrepo log of the container step git source level error ts caller git git go msg error running git exit status nwarning redirecting to couldn t find remote ref warning n stacktrace github com tektoncd pipeline pkg git run n tgithub com tektoncd pipeline pkg git git go ngithub com tektoncd pipeline pkg git fetch n tgithub com tektoncd pipeline pkg git git go nmain main n tgithub com tektoncd pipeline cmd git init main go nruntime main n truntime proc go level fatal ts caller git init main go msg error fetching git repository failed to fetch exit status stacktrace main main n tgithub com tektoncd pipeline cmd git init main go nruntime main n truntime proc go img width alt screen shot at pm src update it turns out this bug happens only in the rancher single install setup | 0 |
169,405 | 20,841,678,385 | IssuesEvent | 2022-03-21 01:17:44 | turkdevops/core | https://api.github.com/repos/turkdevops/core | closed | CVE-2021-35065 (High) detected in glob-parent-2.0.0.tgz - autoclosed | security vulnerability | ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent</p>
<p>
Dependency Hierarchy:
- sdk-1.5.1.tgz (Root Library)
- microbundle-0.3.1.tgz
- rollup-plugin-postcss-1.2.9.tgz
- rollup-pluginutils-2.0.1.tgz
- micromatch-2.3.11.tgz
- parse-glob-3.0.4.tgz
- glob-base-0.3.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gulpjs/glob-parent/pull/49">https://github.com/gulpjs/glob-parent/pull/49</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-35065 (High) detected in glob-parent-2.0.0.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent</p>
<p>
Dependency Hierarchy:
- sdk-1.5.1.tgz (Root Library)
- microbundle-0.3.1.tgz
- rollup-plugin-postcss-1.2.9.tgz
- rollup-pluginutils-2.0.1.tgz
- micromatch-2.3.11.tgz
- parse-glob-3.0.4.tgz
- glob-base-0.3.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gulpjs/glob-parent/pull/49">https://github.com/gulpjs/glob-parent/pull/49</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in glob parent tgz autoclosed cve high severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file package json path to vulnerable library node modules glob parent dependency hierarchy sdk tgz root library microbundle tgz rollup plugin postcss tgz rollup pluginutils tgz micromatch tgz parse glob tgz glob base tgz x glob parent tgz vulnerable library found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource | 0 |
148,846 | 23,390,895,577 | IssuesEvent | 2022-08-11 17:43:38 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | VSCode intellisense — TypeScript type Promise<> displays its contents as if awaited | typescript *as-designed new release | <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: **Yes — Tested with browser incognito at https://vscode.dev**
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.70.1
- OS Version: Windows 11 Pro 64-bit
I must admit that I cannot fathom the core of this problem. A `Promise` in **TypeScript** is displayed in VSCode's intellisense, as if its awaited like seen on these images:
In TypeScript Playground I get following result:

However inside VSCode, the promise is as if its awaited:

I cannot replicate a simple version of my issue, but you can experience same phenomenon by following the reproduction
Because of the `Promise` being displayed correctly in the TypeScript Playground, I add this issue @ VSCode
Steps to Reproduce:
1. Open an incognito browser and visit https://vscode.dev/ and open a new tab
2. Visit my reproduction from [TypeScript Playground](https://www.typescriptlang.org/play?target=99&jsx=0&ssl=192&ssc=1&pln=184&pc=1#code/PQKhAIGUDcFMBsAusDSBLR4TAFA4CawDG8AhgE6zgB2pAtrAM4AOpRVAgs8+AN44BINNWTkAZmyoAZAPZFS8Rn3ABfQcNET24AAplEYmeTrK1QkbHGTd5NNFLIAotWin1Fq9p0BXAEbw0ImdXXlV3TWtIJkY0GWo3cwjtSERvMTE3NRwNSy0qACVYAEdvJkRHOBEAHh0KeiVYAA9kanwlQqIjfCrGRFtqAHMAGnBe-oGAPnAAXnAOrp6+4WHRpcGJkZk6DAApRjiZ8AlFWCn+ARI0WBEOfHxKRkYALlXxwXg5BWfwLmYAOlk8kUglY5HqL1qYLojBB+kMxhehVI+Di8AAnlVfn89A54XQJoJKCUyi8trt9vEmi02uA+qVwAB+cAAeW2iCqhWJvRGAHIAFYUnlTRHFUq9QkybzIACS+BeY2W4AAPjRvPB4IJvOR4C8AKr5KQ4LI4YDALBgcAAWVIA0CeEQaOYVF11Fi1AAKjJpR5GMREG6qrqptNBAAKXXgKnXGmkahoxngUMAaz1AEoZlNoDI0PhwC9qLA4OR01HWkpQ8mXsIxJZwNL09NM9n8OmmdK8zRC5Y8E1mEZMDlPFRIGg6MwAmI0czmP64kpzqAQIIACIyWkACyoMTHE8CDjd4EoRC1MTg6L+ggAAoQJGrMMdfYJsIJCLBmAyXr4ZDJ4LBY0a8FNcAADF9BacBSFpR0qAAdwwddJUwGCjHINFFVIL8pQ3KgHjvC8HSdECwOuKpBA4ON3WgoZBGnWdqAaZpoyUEcdzQSdaLdJRZl4FRqIEYMWRnTiAG0eVfZgeQAXUjRiy1pchSkEJlQmElBYDRSjCOEcAk3UmQMnIjToMkl4WPHNiMUMzTYFU9TrMkkYOLnKYzBeFS1KMrT4l0tF9J+CjjJeKzoNszzYGkrJe37KDCLMicMTIgKnT4pz6Jk6lmNHcz2KEudDh4viBNAhwWkxJLTnSpj-LCpSiJKkjguSwS6MYAkBCC8q8BNM0wAgGAEGQdBMAALUsNcOB0aU8EHPJwGXWBbyQQoWDnKhzl6BxvG+ahvDoXxuwEL98DREVOnIbpYzREZLqmFVLsETdkUsb55nOqpLuuuM2qOtFdV9OVwC-H8-2oQQZBTQHv1-f8BEofA0CPZAAaB6HQYEDbUkYd0qXlNYBkEAjYFxt4BC1HVXmWACcEJqAz2QZa+3o2AqndSq5Nei640+tEBJu-LBFU8BtJ8vz5sWxAGdWkzwBQNmaRFjJWaZd1VOkl4xdIO9JaZ1WqZp7XfQAYQUeBfDYJMWbl9piAWD6IK+w4+dmUMiF8F5QweRnfQhcgyV9Kp+qQWADeZ90JgmBsmxzSPwCzHMuqAgAhc3ox+SaYqYaaPFmibpW9ZgpXnQQfo-e2rsEYlUNL+6siAnRfe2f3QnB92Y9ZlQCRp3OQ8t0saRUnzieWaXQz+MeKAGb5LuEySY8u1QBPrv3Q4JanoLT6UqnzrC+6UXPt8QLi+A7w5OTFcpKnZDnFnGEYFXWEY6QqgAyZQiXPtzwAFOJW4zGxl63tQAuiBRI-Ski5MI9p14AE0A6IE2gxDKqo9qWBGKzXecwbZvTtjdEYssMH30mIcbuTAvbM3cupIW8QUAjzbuAV+oQMZbVMvAzGmwIZP1UAvNehEABycCEFW2QftcgaChHXxwV9PBQjCECRIStJmVQKHxm0jQ3+jZwCswYasBBLCEHsJeA+KgKhuE4E6PRTAzIkzgBsbYux9iHH2NmCzCYoYZC+D5C8d0MdQyMNYcw8AAAmAADMEkYY8-juL5KodMpAlDeGoEmagMgYLxDieAWBISwmaN5ODIUZi5yYENpQBwsBcyOIqU48ALi3EeK8T4vxuignBIAIzhLHlEmJEF4mJOSak7pGSqghLaTk8APIiAlKRvk8xvQfhEHYDOMplTlmHBqVE+pf9fE6Mxi8EJgT2mRI8V09JCSkkpLSUoTJwT9mjJ5GwBZUyCQzMwLwuIHApQIVsKwuwsBvR4n3AcZxYdameM0Q07ZASQkAGYDmdJULEnpZz+npKuTC25yTqDvMQJ8jA+44B-KMHQAF1BpmFPAK8w2cQWiYBWY4oFrj1lgs2Y0nZzSAAssKjnwoGacvpFzBkhI5eimQlKLAiFJRYzBvpECiupbSux9KQUbI0Vsph3wQkAFZOXRO5Sc3p5yBlXK1bch4sAZVUuuIgCVszIT+gULKy18rVnAsZd45lEL1XBIAGzauOYivlhqhnerQbyUEdr4AOvFU8sllo7xoBSIIp1NjFWuvBWq3ZwSADsvrdX+oNSioN2bbl0DjQmzG1rMAcHgCU46hQ+zkCRvKlNdSmUqpZZC4JAAOHNCLwC8vzZcoN3bbkKBrWiOt-YykVrrJaP6Syk0OObaCt1baPW7MCT68AES4W9v7ciwdgSN0hrGaOOd+B8kFMlcnfAZ8yjOoZS2ldUxVX+O+Gy0JPaF02L1Ui-l-D33ZPdLyM2N7RRlAvTTS0ZqEK5mdu-MoFRLVBUmu9L6McOAwVIBgMpmJJo9xuqvGa1hnD4D7BoPgAtB4U0GNLKD2KZD4Cpt1K0pBdIbwTmaQodAZBwCUJdbFipsUOHACiJgJLMBbSoNizgKGbo8KoDoQISY6MwcYByMoucBIqWAgkogdEqE6T0hkZaiBc7SxM7nYSOnqB6bdNJDBHswO9EQyIZDm8bpzzjAmCzk0rO6bomrTsRZIFAQAOJmqUNJ8Alc0TAB+oeUUCNYAMBEEoFgxA2JXFzMIGIhAN5VE7uvA+amOBCNzqhnmciysoerLWYrUw2xAMLh2AsRYuo02NuqM2RALaFFSOQSkskYwO1mH1rUg2kGj3HuQSeLxp6zz-vPJkjmFHe1pgNYOpDVrqf6x6aCO3xvhxjnHXM+YuzkEAmaa0ulIubg3gl3bGcEvHnIKeWA8nMEvbeyphjY2BtVEtEIn7oHdvWTUwR-mAhhJ-fiMLIzVoR6u1Lp1025sAfQ7NeNySEc-4dBPD84HMOAdTFfkvRuzNA70y20zAkJiGGCAACRuUEFDmHBmFYI5eFUfBQ3ywRMunNuMM9PM8xdr4UustX4o+6xbS0GPdvY5jmTtA-sYfWW5+HQQaha6ccxwNvjsOfR+gPH5E2D3xsZyUH5KLPn2wMHo4xoCzcIahjoXT5QIHb29HUVMduH3Cd67206NTgOHP85m1PIXC2VXh9m2XYXi2RvgBdG6T03pRC+ls3ETEmHsPdDV-ty04dV7MdFUWQ+2EpWVvTvb1TVDctUEgn8CvYg0C-ioYgNcUWABeY0ILMDQE9yT2X4hRbEL7MVjuzSOHyuAPsvRXduZDvQjeIcRgDDNa7iJYQgKA9mDyefVrlRjI30fiJ-vWOwHKzPjBJGyMiC0wLQHcPfIZEcCPeD3uoBZXioA4BanIR6AzVnoqhHBhJLRsdhJglscccNFQA8dXsCdoNftA8AdkCQdxswdQDwCYCV9sB-5ycA46ZNtVtTg9Z153R1NehNNDhrsr8UNFMetgc1NbdNdLtzQIBk4etU5c4s4kgFMZBZlzgfpmcBABAGBHgbQiZqNhgWcBhvxkYoYQYtcqYooG1wBnk59BDMBZg4k0QbNExP8L4kN8sdBtDYCzhBBNDQh5CGMRgJDGApCuFdDc9MAjCXNEA-gjC-hv5qBXd1AMhQxrRsUvDYwUQ6BXdwAphgk-gNVUwWdKBHsrFfEWdDoGMToKMxCxCHCpCXgeQAAJBAD4cAAAdSMHgHwAAEIeRUizABB4VCRA9wBr0vdEBfEGjjRL1ZlhBCBGhIdD8mN2Deo5ouwPhmAUt7wJ9qR2MgIABaTRaCPjSgcADfAsMERtD4IEdEcAGCTceIQgM8GQAfQYaiICGxfYbCeMZCcTQGKgKKf6OfPwAIEgNEKBQibGWZbidwXorxKoQmPyHopoWnPATQ0gAffKExdJD4xAPAPQgwsQfzA8MQPw9MfgQQICd0dcFXDQowKgFXRgekGmbEwwBJXMYTAIAscAFpKFTtKwslT2VaR2AfP4QExoP4Q-LZEQ5QWwgGJ+ew6IXIsZOJc9LheFC8QQCUtQIAA)
3. Copy the contents of the playground into the VSCode tab
4. At the bottom of the file (at line 187), add a period and use VSCode intellisense to get results - which should display as the image above
I apologize for the complexity of the situation, for I am not exactly sure what causes the bug. | 1.0 | VSCode intellisense — TypeScript type Promise<> displays its contents as if awaited - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: **Yes — Tested with browser incognito at https://vscode.dev**
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.70.1
- OS Version: Windows 11 Pro 64-bit
I must admit that I cannot fathom the core of this problem. A `Promise` in **TypeScript** is displayed in VSCode's intellisense, as if its awaited like seen on these images:
In TypeScript Playground I get following result:

However inside VSCode, the promise is as if its awaited:

I cannot replicate a simple version of my issue, but you can experience same phenomenon by following the reproduction
Because of the `Promise` being displayed correctly in the TypeScript Playground, I add this issue @ VSCode
Steps to Reproduce:
1. Open an incognito browser and visit https://vscode.dev/ and open a new tab
2. Visit my reproduction from [TypeScript Playground](https://www.typescriptlang.org/play?target=99&jsx=0&ssl=192&ssc=1&pln=184&pc=1#code/PQKhAIGUDcFMBsAusDSBLR4TAFA4CawDG8AhgE6zgB2pAtrAM4AOpRVAgs8+AN44BINNWTkAZmyoAZAPZFS8Rn3ABfQcNET24AAplEYmeTrK1QkbHGTd5NNFLIAotWin1Fq9p0BXAEbw0ImdXXlV3TWtIJkY0GWo3cwjtSERvMTE3NRwNSy0qACVYAEdvJkRHOBEAHh0KeiVYAA9kanwlQqIjfCrGRFtqAHMAGnBe-oGAPnAAXnAOrp6+4WHRpcGJkZk6DAApRjiZ8AlFWCn+ARI0WBEOfHxKRkYALlXxwXg5BWfwLmYAOlk8kUglY5HqL1qYLojBB+kMxhehVI+Di8AAnlVfn89A54XQJoJKCUyi8trt9vEmi02uA+qVwAB+cAAeW2iCqhWJvRGAHIAFYUnlTRHFUq9QkybzIACS+BeY2W4AAPjRvPB4IJvOR4C8AKr5KQ4LI4YDALBgcAAWVIA0CeEQaOYVF11Fi1AAKjJpR5GMREG6qrqptNBAAKXXgKnXGmkahoxngUMAaz1AEoZlNoDI0PhwC9qLA4OR01HWkpQ8mXsIxJZwNL09NM9n8OmmdK8zRC5Y8E1mEZMDlPFRIGg6MwAmI0czmP64kpzqAQIIACIyWkACyoMTHE8CDjd4EoRC1MTg6L+ggAAoQJGrMMdfYJsIJCLBmAyXr4ZDJ4LBY0a8FNcAADF9BacBSFpR0qAAdwwddJUwGCjHINFFVIL8pQ3KgHjvC8HSdECwOuKpBA4ON3WgoZBGnWdqAaZpoyUEcdzQSdaLdJRZl4FRqIEYMWRnTiAG0eVfZgeQAXUjRiy1pchSkEJlQmElBYDRSjCOEcAk3UmQMnIjToMkl4WPHNiMUMzTYFU9TrMkkYOLnKYzBeFS1KMrT4l0tF9J+CjjJeKzoNszzYGkrJe37KDCLMicMTIgKnT4pz6Jk6lmNHcz2KEudDh4viBNAhwWkxJLTnSpj-LCpSiJKkjguSwS6MYAkBCC8q8BNM0wAgGAEGQdBMAALUsNcOB0aU8EHPJwGXWBbyQQoWDnKhzl6BxvG+ahvDoXxuwEL98DREVOnIbpYzREZLqmFVLsETdkUsb55nOqpLuuuM2qOtFdV9OVwC-H8-2oQQZBTQHv1-f8BEofA0CPZAAaB6HQYEDbUkYd0qXlNYBkEAjYFxt4BC1HVXmWACcEJqAz2QZa+3o2AqndSq5Nei640+tEBJu-LBFU8BtJ8vz5sWxAGdWkzwBQNmaRFjJWaZd1VOkl4xdIO9JaZ1WqZp7XfQAYQUeBfDYJMWbl9piAWD6IK+w4+dmUMiF8F5QweRnfQhcgyV9Kp+qQWADeZ90JgmBsmxzSPwCzHMuqAgAhc3ox+SaYqYaaPFmibpW9ZgpXnQQfo-e2rsEYlUNL+6siAnRfe2f3QnB92Y9ZlQCRp3OQ8t0saRUnzieWaXQz+MeKAGb5LuEySY8u1QBPrv3Q4JanoLT6UqnzrC+6UXPt8QLi+A7w5OTFcpKnZDnFnGEYFXWEY6QqgAyZQiXPtzwAFOJW4zGxl63tQAuiBRI-Ski5MI9p14AE0A6IE2gxDKqo9qWBGKzXecwbZvTtjdEYssMH30mIcbuTAvbM3cupIW8QUAjzbuAV+oQMZbVMvAzGmwIZP1UAvNehEABycCEFW2QftcgaChHXxwV9PBQjCECRIStJmVQKHxm0jQ3+jZwCswYasBBLCEHsJeA+KgKhuE4E6PRTAzIkzgBsbYux9iHH2NmCzCYoYZC+D5C8d0MdQyMNYcw8AAAmAADMEkYY8-juL5KodMpAlDeGoEmagMgYLxDieAWBISwmaN5ODIUZi5yYENpQBwsBcyOIqU48ALi3EeK8T4vxuignBIAIzhLHlEmJEF4mJOSak7pGSqghLaTk8APIiAlKRvk8xvQfhEHYDOMplTlmHBqVE+pf9fE6Mxi8EJgT2mRI8V09JCSkkpLSUoTJwT9mjJ5GwBZUyCQzMwLwuIHApQIVsKwuwsBvR4n3AcZxYdameM0Q07ZASQkAGYDmdJULEnpZz+npKuTC25yTqDvMQJ8jA+44B-KMHQAF1BpmFPAK8w2cQWiYBWY4oFrj1lgs2Y0nZzSAAssKjnwoGacvpFzBkhI5eimQlKLAiFJRYzBvpECiupbSux9KQUbI0Vsph3wQkAFZOXRO5Sc3p5yBlXK1bch4sAZVUuuIgCVszIT+gULKy18rVnAsZd45lEL1XBIAGzauOYivlhqhnerQbyUEdr4AOvFU8sllo7xoBSIIp1NjFWuvBWq3ZwSADsvrdX+oNSioN2bbl0DjQmzG1rMAcHgCU46hQ+zkCRvKlNdSmUqpZZC4JAAOHNCLwC8vzZcoN3bbkKBrWiOt-YykVrrJaP6Syk0OObaCt1baPW7MCT68AES4W9v7ciwdgSN0hrGaOOd+B8kFMlcnfAZ8yjOoZS2ldUxVX+O+Gy0JPaF02L1Ui-l-D33ZPdLyM2N7RRlAvTTS0ZqEK5mdu-MoFRLVBUmu9L6McOAwVIBgMpmJJo9xuqvGa1hnD4D7BoPgAtB4U0GNLKD2KZD4Cpt1K0pBdIbwTmaQodAZBwCUJdbFipsUOHACiJgJLMBbSoNizgKGbo8KoDoQISY6MwcYByMoucBIqWAgkogdEqE6T0hkZaiBc7SxM7nYSOnqB6bdNJDBHswO9EQyIZDm8bpzzjAmCzk0rO6bomrTsRZIFAQAOJmqUNJ8Alc0TAB+oeUUCNYAMBEEoFgxA2JXFzMIGIhAN5VE7uvA+amOBCNzqhnmciysoerLWYrUw2xAMLh2AsRYuo02NuqM2RALaFFSOQSkskYwO1mH1rUg2kGj3HuQSeLxp6zz-vPJkjmFHe1pgNYOpDVrqf6x6aCO3xvhxjnHXM+YuzkEAmaa0ulIubg3gl3bGcEvHnIKeWA8nMEvbeyphjY2BtVEtEIn7oHdvWTUwR-mAhhJ-fiMLIzVoR6u1Lp1025sAfQ7NeNySEc-4dBPD84HMOAdTFfkvRuzNA70y20zAkJiGGCAACRuUEFDmHBmFYI5eFUfBQ3ywRMunNuMM9PM8xdr4UustX4o+6xbS0GPdvY5jmTtA-sYfWW5+HQQaha6ccxwNvjsOfR+gPH5E2D3xsZyUH5KLPn2wMHo4xoCzcIahjoXT5QIHb29HUVMduH3Cd67206NTgOHP85m1PIXC2VXh9m2XYXi2RvgBdG6T03pRC+ls3ETEmHsPdDV-ty04dV7MdFUWQ+2EpWVvTvb1TVDctUEgn8CvYg0C-ioYgNcUWABeY0ILMDQE9yT2X4hRbEL7MVjuzSOHyuAPsvRXduZDvQjeIcRgDDNa7iJYQgKA9mDyefVrlRjI30fiJ-vWOwHKzPjBJGyMiC0wLQHcPfIZEcCPeD3uoBZXioA4BanIR6AzVnoqhHBhJLRsdhJglscccNFQA8dXsCdoNftA8AdkCQdxswdQDwCYCV9sB-5ycA46ZNtVtTg9Z153R1NehNNDhrsr8UNFMetgc1NbdNdLtzQIBk4etU5c4s4kgFMZBZlzgfpmcBABAGBHgbQiZqNhgWcBhvxkYoYQYtcqYooG1wBnk59BDMBZg4k0QbNExP8L4kN8sdBtDYCzhBBNDQh5CGMRgJDGApCuFdDc9MAjCXNEA-gjC-hv5qBXd1AMhQxrRsUvDYwUQ6BXdwAphgk-gNVUwWdKBHsrFfEWdDoGMToKMxCxCHCpCXgeQAAJBAD4cAAAdSMHgHwAAEIeRUizABB4VCRA9wBr0vdEBfEGjjRL1ZlhBCBGhIdD8mN2Deo5ouwPhmAUt7wJ9qR2MgIABaTRaCPjSgcADfAsMERtD4IEdEcAGCTceIQgM8GQAfQYaiICGxfYbCeMZCcTQGKgKKf6OfPwAIEgNEKBQibGWZbidwXorxKoQmPyHopoWnPATQ0gAffKExdJD4xAPAPQgwsQfzA8MQPw9MfgQQICd0dcFXDQowKgFXRgekGmbEwwBJXMYTAIAscAFpKFTtKwslT2VaR2AfP4QExoP4Q-LZEQ5QWwgGJ+ew6IXIsZOJc9LheFC8QQCUtQIAA)
3. Copy the contents of the playground into the VSCode tab
4. At the bottom of the file (at line 187), add a period and use VSCode intellisense to get results - which should display as the image above
I apologize for the complexity of the situation, for I am not exactly sure what causes the bug. | non_priority | vscode intellisense — typescript type promise displays its contents as if awaited does this issue occur when all extensions are disabled yes — tested with browser incognito at report issue dialog can assist with this vs code version os version windows pro bit i must admit that i cannot fathom the core of this problem a promise in typescript is displayed in vscode s intellisense as if its awaited like seen on these images in typescript playground i get following result however inside vscode the promise is as if its awaited i cannot replicate a simple version of my issue but you can experience same phenomenon by following the reproduction because of the promise being displayed correctly in the typescript playground i add this issue vscode steps to reproduce open an incognito browser and visit and open a new tab visit my reproduction from copy the contents of the playground into the vscode tab at the bottom of the file at line add a period and use vscode intellisense to get results which should display as the image above i apologize for the complexity of the situation for i am not exactly sure what causes the bug | 0 |
130,379 | 27,660,002,415 | IssuesEvent | 2023-03-12 12:14:36 | conbench/conbench | https://api.github.com/repos/conbench/conbench | opened | benchmark result view: clarify time axis on plot | UI/UX time evolution & code evolution | This is a small task: in the benchmark result view, make it precise and explicit what the time axis shows. | 1.0 | benchmark result view: clarify time axis on plot - This is a small task: in the benchmark result view, make it precise and explicit what the time axis shows. | non_priority | benchmark result view clarify time axis on plot this is a small task in the benchmark result view make it precise and explicit what the time axis shows | 0 |
45,169 | 18,444,609,115 | IssuesEvent | 2021-10-14 23:07:07 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Does Autoscale allows "scale up/down" of Azure App Service? | app-service/svc triaged cxp doc-enhancement Pri1 |
[Here it's stated ](https://docs.microsoft.com/en-us/azure/app-service/manage-scale-up)
"Scale up: Get more CPU, memory, disk space, and extra features like dedicated virtual machines (VMs), custom domains and certificates, staging slots, **_autoscaling_**, and more. You scale up by changing the pricing tier of the App Service plan that your app belongs to."
However, [Azure Monitor Autoscaling indicates](https://docs.microsoft.com/en-us/azure/azure-monitor/autoscale/autoscale-overview#horizontal-vs-vertical-scaling) that "_**Autoscale only scales horizontally**_"
and [horizontal scaling is only for Virtual Machine Scale Sets](https://docs.microsoft.com/en-us/azure/virtual-machine-scale-sets/virtual-machine-scale-sets-vertical-scale-reprovision)
So, the word "autoscaling" from scale up definition should be removed. If not, please pardon me that my understanding is incorrect.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 44c8de6c-a71b-7c7d-69f5-b1ec446f33ff
* Version Independent ID: 85f9c74a-7039-f466-f2da-bef5eaf106a1
* Content: [Scale up features and capacities - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/manage-scale-up)
* Content Source: [articles/app-service/manage-scale-up.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/app-service/manage-scale-up.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | 1.0 | Does Autoscale allows "scale up/down" of Azure App Service? -
[Here it's stated ](https://docs.microsoft.com/en-us/azure/app-service/manage-scale-up)
"Scale up: Get more CPU, memory, disk space, and extra features like dedicated virtual machines (VMs), custom domains and certificates, staging slots, **_autoscaling_**, and more. You scale up by changing the pricing tier of the App Service plan that your app belongs to."
However, [Azure Monitor Autoscaling indicates](https://docs.microsoft.com/en-us/azure/azure-monitor/autoscale/autoscale-overview#horizontal-vs-vertical-scaling) that "_**Autoscale only scales horizontally**_"
and [horizontal scaling is only for Virtual Machine Scale Sets](https://docs.microsoft.com/en-us/azure/virtual-machine-scale-sets/virtual-machine-scale-sets-vertical-scale-reprovision)
So, the word "autoscaling" from scale up definition should be removed. If not, please pardon me that my understanding is incorrect.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 44c8de6c-a71b-7c7d-69f5-b1ec446f33ff
* Version Independent ID: 85f9c74a-7039-f466-f2da-bef5eaf106a1
* Content: [Scale up features and capacities - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/manage-scale-up)
* Content Source: [articles/app-service/manage-scale-up.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/app-service/manage-scale-up.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | non_priority | does autoscale allows scale up down of azure app service scale up get more cpu memory disk space and extra features like dedicated virtual machines vms custom domains and certificates staging slots autoscaling and more you scale up by changing the pricing tier of the app service plan that your app belongs to however that autoscale only scales horizontally and so the word autoscaling from scale up definition should be removed if not please pardon me that my understanding is incorrect document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login cephalin microsoft alias cephalin | 0 |
317,518 | 27,243,014,122 | IssuesEvent | 2023-02-21 22:20:44 | microsoft/vscode-remote-release | https://api.github.com/repos/microsoft/vscode-remote-release | closed | Test: Download WSL server build in the background | testplan-item | Refs: https://github.com/microsoft/vscode-remote-release/issues/3839
- [x] Windows @connor4312 with WSL
- [x] Windows @luabud with WSL
Complexity: 3
[Create Issue](https://github.com/microsoft/vscode-remote-release/issues/new?body=Testing+%238045%0A%0A&assignees=aeschli)
-----------
Prepare:
- Install the latest version of the WSL extension (v0.75.3)
- Set `"remote.WSL.downloadInBackground": true
- Open the WSL log (Command: WSL: Show Log...)
- Verify that there is a message `Download in background is enabled`
- Close VS Code and install a previous VS Code build
- Check the WSL Log that you see again the `Download in background is enabled` message
- At some point VSCode should detect an update and render a decorator over the settings gear. Do not update!
- After maximum 2 minutes the WSL log should also detect the new update and print a message `New commit detected: ${commitId}`
- if you have not used WSL since the update of the WSL extension you will see a message `No recently used WSL platforms found. Skipping download.`
- open one or more WSL window of different distros, e.g. Ubuntu and Alpine
- After maximum 2 minutes the WSL log check again for downloads and download the server builds | 1.0 | Test: Download WSL server build in the background - Refs: https://github.com/microsoft/vscode-remote-release/issues/3839
- [x] Windows @connor4312 with WSL
- [x] Windows @luabud with WSL
Complexity: 3
[Create Issue](https://github.com/microsoft/vscode-remote-release/issues/new?body=Testing+%238045%0A%0A&assignees=aeschli)
-----------
Prepare:
- Install the latest version of the WSL extension (v0.75.3)
- Set `"remote.WSL.downloadInBackground": true
- Open the WSL log (Command: WSL: Show Log...)
- Verify that there is a message `Download in background is enabled`
- Close VS Code and install a previous VS Code build
- Check the WSL Log that you see again the `Download in background is enabled` message
- At some point VSCode should detect an update and render a decorator over the settings gear. Do not update!
- After maximum 2 minutes the WSL log should also detect the new update and print a message `New commit detected: ${commitId}`
- if you have not used WSL since the update of the WSL extension you will see a message `No recently used WSL platforms found. Skipping download.`
- open one or more WSL window of different distros, e.g. Ubuntu and Alpine
- After maximum 2 minutes the WSL log check again for downloads and download the server builds | non_priority | test download wsl server build in the background refs windows with wsl windows luabud with wsl complexity prepare install the latest version of the wsl extension set remote wsl downloadinbackground true open the wsl log command wsl show log verify that there is a message download in background is enabled close vs code and install a previous vs code build check the wsl log that you see again the download in background is enabled message at some point vscode should detect an update and render a decorator over the settings gear do not update after maximum minutes the wsl log should also detect the new update and print a message new commit detected commitid if you have not used wsl since the update of the wsl extension you will see a message no recently used wsl platforms found skipping download open one or more wsl window of different distros e g ubuntu and alpine after maximum minutes the wsl log check again for downloads and download the server builds | 0 |
307,860 | 23,219,820,289 | IssuesEvent | 2022-08-02 17:04:57 | workday-group-3/referee-handbook-Frontend | https://api.github.com/repos/workday-group-3/referee-handbook-Frontend | closed | Update API docs to match current backend | documentation | - [x] Update all endpoints
- [x] Update variables according to sql database
- [x] Add rest api request endpoints | 1.0 | Update API docs to match current backend - - [x] Update all endpoints
- [x] Update variables according to sql database
- [x] Add rest api request endpoints | non_priority | update api docs to match current backend update all endpoints update variables according to sql database add rest api request endpoints | 0 |
49,328 | 6,197,051,967 | IssuesEvent | 2017-07-05 15:57:01 | nutofem/nuto | https://api.github.com/repos/nutofem/nuto | opened | Refactor NodeBase and NodeDof | cleanup design | **Problem:**
- NodeBase is an abstract class that only has one child.
- Many functions throw an exception instead of being pure virtual
**Reason:**
- NodeDof used to be templated over various DOFs
**Solution:**
- Merge NodeBase and NodeDof into a single class to improve readability.
| 1.0 | Refactor NodeBase and NodeDof - **Problem:**
- NodeBase is an abstract class that only has one child.
- Many functions throw an exception instead of being pure virtual
**Reason:**
- NodeDof used to be templated over various DOFs
**Solution:**
- Merge NodeBase and NodeDof into a single class to improve readability.
| non_priority | refactor nodebase and nodedof problem nodebase is an abstract class that only has one child many functions throw an exception instead of being pure virtual reason nodedof used to be templated over various dofs solution merge nodebase and nodedof into a single class to improve readability | 0 |
75,219 | 25,594,851,249 | IssuesEvent | 2022-12-01 15:29:09 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Field.contains(String[]) does not work for text[] fields | T: Defect C: DB: PostgreSQL P: Medium E: All Editions | Currently it converts String[] to varchar[] and it results into `ERROR: operator does not exist: text[] @> character varying[]`. As a result, you have to write something like `DSL.array(DSL.val(str).cast(PostgresDataType.TEXT))`.
P.S. while examining the issue I found that the problem lies not in `contains()` implementation itself, but somewhere deeper in type handling code.
| 1.0 | Field.contains(String[]) does not work for text[] fields - Currently it converts String[] to varchar[] and it results into `ERROR: operator does not exist: text[] @> character varying[]`. As a result, you have to write something like `DSL.array(DSL.val(str).cast(PostgresDataType.TEXT))`.
P.S. while examining the issue I found that the problem lies not in `contains()` implementation itself, but somewhere deeper in type handling code.
| non_priority | field contains string does not work for text fields currently it converts string to varchar and it results into error operator does not exist text character varying as a result you have to write something like dsl array dsl val str cast postgresdatatype text p s while examining the issue i found that the problem lies not in contains implementation itself but somewhere deeper in type handling code | 0 |
82,706 | 10,308,498,293 | IssuesEvent | 2019-08-29 11:05:55 | ComputationalRadiationPhysics/picongpu | https://api.github.com/repos/ComputationalRadiationPhysics/picongpu | opened | add documentation radiation plugin python module | component: plugin documentation | We have a python module for reading hdf5 data from the radiation plugin. However, this is neither mentioned nor documented in the documentation.
A short python example should be included. | 1.0 | add documentation radiation plugin python module - We have a python module for reading hdf5 data from the radiation plugin. However, this is neither mentioned nor documented in the documentation.
A short python example should be included. | non_priority | add documentation radiation plugin python module we have a python module for reading data from the radiation plugin however this is neither mentioned nor documented in the documentation a short python example should be included | 0 |
38,555 | 8,894,979,285 | IssuesEvent | 2019-01-16 06:57:23 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | closed | Stats: Map Gives "Not a Valid Colour String" Error on Some Devices | Color Schemes [Pri] High [Type] Defect | <!-- Thanks for contributing to Calypso! Pick a clear title ("Editor: add spell check") and proceed. -->
#### Steps to reproduce
1. Starting at URL: https://wordpress.com/stats
2. Try viewing the map
#### What I expected
The map and graph to appear.
#### What happened instead
> You can see that the daily stats do not show up on the timeline. (before I had 2 bars — views & visitors — for each day, or week…). There is no map and there is no colour (before yellow) vertical bar in front of today post (30 years ago — former…)
Several issues with stats not loading, and most notably an error which says "is not a valid colour string".

From the reports so far, it looks like that both issues either occur at once, or none occur at all.
#### Browser / OS version
So far, most the reports I've seen have either not given their browser or their OS. However, the user who took the screenshot above said they use Windows 7 and Internet Explorer 11. One other user said they're using Internet Explorer too.
#### Context / Source
I have not been able to reproduce this successfully, however I've seen several reports on it: https://en.forums.wordpress.com/topic/dashboard-and-statistics/
https://en.forums.wordpress.com/topic/none-valid-colour-string-on-stats-page/
https://en.forums.wordpress.com/topic/stats-345/
https://en.forums.wordpress.com/topic/random-color-change-to-a-lot-of-things/page/2/#post-3248000
https://en.forums.wordpress.com/topic/changing-colors-lost-functionality/
(cc @flootr, this might be related to #30078 since this issue wasn't reported before then; the error is the same as the one in #30066) | 1.0 | Stats: Map Gives "Not a Valid Colour String" Error on Some Devices - <!-- Thanks for contributing to Calypso! Pick a clear title ("Editor: add spell check") and proceed. -->
#### Steps to reproduce
1. Starting at URL: https://wordpress.com/stats
2. Try viewing the map
#### What I expected
The map and graph to appear.
#### What happened instead
> You can see that the daily stats do not show up on the timeline. (before I had 2 bars — views & visitors — for each day, or week…). There is no map and there is no colour (before yellow) vertical bar in front of today post (30 years ago — former…)
Several issues with stats not loading, and most notably an error which says "is not a valid colour string".

From the reports so far, it looks like that both issues either occur at once, or none occur at all.
#### Browser / OS version
So far, most the reports I've seen have either not given their browser or their OS. However, the user who took the screenshot above said they use Windows 7 and Internet Explorer 11. One other user said they're using Internet Explorer too.
#### Context / Source
I have not been able to reproduce this successfully, however I've seen several reports on it: https://en.forums.wordpress.com/topic/dashboard-and-statistics/
https://en.forums.wordpress.com/topic/none-valid-colour-string-on-stats-page/
https://en.forums.wordpress.com/topic/stats-345/
https://en.forums.wordpress.com/topic/random-color-change-to-a-lot-of-things/page/2/#post-3248000
https://en.forums.wordpress.com/topic/changing-colors-lost-functionality/
(cc @flootr, this might be related to #30078 since this issue wasn't reported before then; the error is the same as the one in #30066) | non_priority | stats map gives not a valid colour string error on some devices steps to reproduce starting at url try viewing the map what i expected the map and graph to appear what happened instead you can see that the daily stats do not show up on the timeline before i had bars — views visitors — for each day or week… there is no map and there is no colour before yellow vertical bar in front of today post years ago — former… several issues with stats not loading and most notably an error which says is not a valid colour string from the reports so far it looks like that both issues either occur at once or none occur at all browser os version so far most the reports i ve seen have either not given their browser or their os however the user who took the screenshot above said they use windows and internet explorer one other user said they re using internet explorer too context source i have not been able to reproduce this successfully however i ve seen several reports on it cc flootr this might be related to since this issue wasn t reported before then the error is the same as the one in | 0 |
11,920 | 2,668,989,976 | IssuesEvent | 2015-03-23 13:03:33 | contao/core-bundle | https://api.github.com/repos/contao/core-bundle | closed | Return value of `$kernel->getContaoBundles()` not recognized | defect | > <a href="https://github.com/leofeyer"><img src="https://avatars.githubusercontent.com/u/1192057?v=3" align="left" width="42" height="42" hspace="10"></img></a> [Issue](https://github.com/contao/contao/issues/48) by @leofeyer
Friday Feb 13, 2015 at 11:34 GMT
@aschempp Any idea why the return value of `$kernel->getContaoBundles()` is not recognized by PhpStorm? I had to add an inline `@var` comment:
https://github.com/contao/core-bundle/blob/develop/contao/library/Contao/Automator.php#L556
| 1.0 | Return value of `$kernel->getContaoBundles()` not recognized - > <a href="https://github.com/leofeyer"><img src="https://avatars.githubusercontent.com/u/1192057?v=3" align="left" width="42" height="42" hspace="10"></img></a> [Issue](https://github.com/contao/contao/issues/48) by @leofeyer
Friday Feb 13, 2015 at 11:34 GMT
@aschempp Any idea why the return value of `$kernel->getContaoBundles()` is not recognized by PhpStorm? I had to add an inline `@var` comment:
https://github.com/contao/core-bundle/blob/develop/contao/library/Contao/Automator.php#L556
| non_priority | return value of kernel getcontaobundles not recognized by leofeyer friday feb at gmt aschempp any idea why the return value of kernel getcontaobundles is not recognized by phpstorm i had to add an inline var comment | 0 |
58,253 | 11,856,035,869 | IssuesEvent | 2020-03-25 06:22:15 | NETMouse-projects/ABCNET | https://api.github.com/repos/NETMouse-projects/ABCNET | closed | Функции чтения с клавиатуры для стандартных типов NET | area-code enhancement importance-low namespace-Utils platform-NET type-new functionality v0.5.5 ➥ STALE | **Описание:**
Функции чтения с клавиатуры для всех базовых типов `NET`, а также и для `BigInteger`.
**Способ реализации:**
не предоставляется
**Альтернативные решения:**
не предоставляются | 1.0 | Функции чтения с клавиатуры для стандартных типов NET - **Описание:**
Функции чтения с клавиатуры для всех базовых типов `NET`, а также и для `BigInteger`.
**Способ реализации:**
не предоставляется
**Альтернативные решения:**
не предоставляются | non_priority | функции чтения с клавиатуры для стандартных типов net описание функции чтения с клавиатуры для всех базовых типов net а также и для biginteger способ реализации не предоставляется альтернативные решения не предоставляются | 0 |
128,326 | 10,525,142,713 | IssuesEvent | 2019-09-30 14:34:24 | int-brain-lab/iblrig | https://api.github.com/repos/int-brain-lab/iblrig | closed | Frame2ttl unreliable data when wheel moves | 00 triage 01 wip 02 testing bug help wanted | **Describe the bug**
Accordingly to Nicco, the frame2ttl does not output the expected trace when the wheel moved.
We need to test for the reliability of the frame2ttl separately from the task software.
**To Reproduce**
Record in rig.
**Expected behavior**
The frame2ttl should perfectly follow the synch square signal.
**urgent need**
To test this, have a Bonsai workflow that switches the synch square (or even the whole screen if less complex) at the screen refresh rate (60Hz), for a set number of times (1000 for example).
If the Frame2ttl data has the expected number of switches, we can state with confidence the hardware is not faulty. | 1.0 | Frame2ttl unreliable data when wheel moves - **Describe the bug**
Accordingly to Nicco, the frame2ttl does not output the expected trace when the wheel moved.
We need to test for the reliability of the frame2ttl separately from the task software.
**To Reproduce**
Record in rig.
**Expected behavior**
The frame2ttl should perfectly follow the synch square signal.
**urgent need**
To test this, have a Bonsai workflow that switches the synch square (or even the whole screen if less complex) at the screen refresh rate (60Hz), for a set number of times (1000 for example).
If the Frame2ttl data has the expected number of switches, we can state with confidence the hardware is not faulty. | non_priority | unreliable data when wheel moves describe the bug accordingly to nicco the does not output the expected trace when the wheel moved we need to test for the reliability of the separately from the task software to reproduce record in rig expected behavior the should perfectly follow the synch square signal urgent need to test this have a bonsai workflow that switches the synch square or even the whole screen if less complex at the screen refresh rate for a set number of times for example if the data has the expected number of switches we can state with confidence the hardware is not faulty | 0 |
331,055 | 28,503,993,390 | IssuesEvent | 2023-04-18 19:43:57 | eclipse-openj9/openj9 | https://api.github.com/repos/eclipse-openj9/openj9 | opened | JDK21 testSCCMLTests1_openj9_1_FAILED - lambda classes | comp:vm test failure jdk21 | Failure link
------------
From [an internal grinder](https://openj9-jenkins.osuosl.org/job/Grinder/2261):
```
15:19:40 Eclipse OpenJ9 VM (build HEAD-1f25b601986, JRE 21 Linux amd64-64-Bit Compressed References 20230322_60 (JIT enabled, AOT enabled)
15:19:40 OpenJ9 - 1f25b601986
15:19:40 OMR - c766c4155b0
15:19:40 JCL - 732652e2726 based on jdk-21+13)
```
[Rerun in Grinder](https://openj9-jenkins.osuosl.org/job/Grinder/parambuild/?SDK_RESOURCE=customized&DOCKER_REQUIRED=false&ACTIVE_NODE_TIMEOUT=&EXTRA_DOCKER_ARGS=&PLATFORM=x86-64_linux&PERSONAL_BUILD=false&DOCKER_REGISTRY_DIR=&ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&SETUP_JCK_RUN=false&DOCKER_REGISTRY_URL_CREDENTIAL_ID=&EXTRA_OPTIONS=&BUILD_IDENTIFIER=&JENKINS_KEY=&LIGHT_WEIGHT_CHECKOUT=false&ARTIFACTORY_SERVER=ci-eclipse-openj9&USE_JRE=false&KEEP_WORKSPACE=false&JDK_VERSION=next&DOCKER_REGISTRY_URL=&EXIT_FAILURE=false&EXIT_SUCCESS=false&JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk.git&JCK_GIT_BRANCH=master&OPENJ9_BRANCH=master&OPENJ9_SHA=origin%2Fpr%2F17201%2Fmerge&VENDOR_TEST_BRANCHES=&OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&JDK_BRANCH=openj9&LABEL_ADDITION=%21sw.os.cent.6&ARTIFACTORY_REPO=ci-openj9&UPSTREAM_TEST_JOB_NUMBER=2261&DOCKERIMAGE_TAG=&RELATED_NODES=&ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&APPLICATION_OPTIONS=&CUSTOMIZED_SDK_URL_CREDENTIAL_ID=e047f50a-26c4-4b82-8525-6da17ec68450&OPENJDK_SHA=&BUILD_LIST=functional&USE_TESTENV_PROPERTIES=false&UPSTREAM_JOB_NUMBER=&STF_OWNER_BRANCH=adoptium%3Amaster&JDK_VENDOR=&BUILDS_TO_KEEP=60&TIME_LIMIT=10&PARALLEL=None&TARGET=testSCCMLTests1_openj9_1&TEST_FLAG=&UPSTREAM_TEST_JOB_NAME=Grinder&VENDOR_TEST_DIRS=&TKG_OWNER_BRANCH=adoptium%3Amaster&OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&GENERATE_JOBS=true&KEEP_REPORTDIR=false&RERUN_ITERATIONS=0&LABEL=&CUSTOMIZED_SDK_URL=https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDKnext_x86-64_linux_Personal%2F60%2FOpenJ9-JDKnext-x86-64_linux-20230322-190845.tar.gz&DAYS_TO_KEEP=7&ADOPTOPENJDK_BRANCH=master&USER_CREDENTIALS_ID=&ITERATIONS=10&VENDOR_TEST_REPOS=&JCK_GIT_REPO=&UPSTREAM_JOB_NAME=&CLOUD_PROVIDER=&CUSTOM_TARGET=&DEBUG_IMAGES_REQUIRED=false&VENDOR_TEST_SHAS=&TEST_IMAGES_REQUIRED=true&ARTIFACTORY_ROOT_DIR=Test&JDK_IMPL=openj9&TEST_TIME=&SSH_AGENT_CREDENTIAL=&AUTO_DETECT=true&SLACK_CHANNEL=&DYNAMIC_COMPILE=true&TKG_ITERATIONS=1&ARCHIVE_TEST_RESULTS=false&NUM_MACHINES=&TRSS_URL=&ADDITIONAL_ARTIFACTS_REQUIRED=&JVM_OPTIONS=) - Change TARGET to run only the failed test targets.
Optional info
-------------
Failure output (captured from console output)
---------------------------------------------
```
15:31:37 ===============================================
15:31:37 Running test testSCCMLTests1_openj9_1 ...
15:31:37 ===============================================
15:31:37 testSCCMLTests1_openj9_1 Start Time: Tue Apr 18 15:31:36 2023 Epoch Time (ms): 1681846296459
15:31:37 variation: Mode610
15:31:37 JVM_OPTIONS: -Xcompressedrefs -Xjit -Xgcpolicy:gencon
15:32:21 Testing: Test 57-a: Make sure that lambda classes work and get stored in the shared class cache
15:32:21 Test start time: 2023/04/18 15:32:20 Eastern Standard Time
15:32:21 Running command: "/home/jenkins/workspace/Grinder_iteration_0/openjdkbinary/j2sdk-image/bin/java" -Xcompressedrefs -Xjit -Xgcpolicy:gencon -Xshareclasses:name=ShareClassesCMLTests -Xtrace:print={j9shr.2259} -cp /home/jenkins/workspace/Grinder_iteration_0/aqa-tests/TKG/../../jvmtest/functional/cmdLineTests/utils/utils.jar org.openj9.test.lambdatests.Test1 0
15:32:21 Time spent starting: 3 milliseconds
15:32:22 Time spent executing: 912 milliseconds
15:32:22 Test result: FAILED
15:32:23 [OUT] 364
15:32:23 [OUT] 1720
15:32:23 [OUT] Lambda test done!
15:32:23 [ERR] 19:32:21.374*0x17000 j9shr.2259 > API j9shr_classStoreTransaction_createSharedClass : enter (classname=com/ibm/sharedclasses/spi/SharedClassProvider romClassSizeFullSize=1416 romClassMinimalSize=1416 lineNumberTableSize=0 localVariableTableSize=0
15:32:23 [ERR] 19:32:21.860 0x17000 j9shr.2259 > API j9shr_classStoreTransaction_createSharedClass : enter (classname=org/openj9/test/lambdatests/Test1$$Lambda/0x0000000000000000 romClassSizeFullSize=592 romClassMinimalSize=592 lineNumberTableSize=0 localVariableTableSize=0
15:32:23 >> Success condition was not found: [Output match: j9shr[\.]2259 > API j9shr_classStoreTransaction_createSharedClass : enter [\(]classname=org/openj9/test/lambdatests/Test1..Lambda.([\d]+)/(0x00000000|0x0000000000000000)]
15:32:23 >> Required condition was found: [Output match: Lambda test done!]
15:32:23 >> Required condition was not found: [Output match: j9shr[\.]2259 > API j9shr_classStoreTransaction_createSharedClass : enter [\(]classname=org/openj9/test/lambdatests/Test1..Lambda.([\d]+)/(0x00000000|0x0000000000000000)]
15:32:23 >> Failure condition was not found: [Output match: Unhandled Exception]
15:32:23 >> Failure condition was not found: [Output match: corrupt]
15:32:23 >> Failure condition was not found: [Output match: Processing dump event]
15:33:05 ---TEST RESULTS---
15:33:05 Number of PASSED tests: 79 out of 87
15:33:05 Number of FAILED tests: 8 out of 87
15:33:05
15:33:05 ---SUMMARY OF FAILED TESTS---
15:33:05 Test 57-a: Make sure that lambda classes work and get stored in the shared class cache
15:33:05 Test 57-b: Make sure that lambda classes are stored in the cache as orphans
15:33:05 Test 57-c: Make sure that when the program runs again lambda classes are used from the cache and not stored again
15:33:05 Test 58: Make sure that when the program runs again without the first lambda class, the second lambda class is still used from the cache and not stored again
15:33:05 Test 59-a: Run and store 10 lambda classes in the cache
15:33:05 Test 59-b: Do not run the first lambda class to check if a class with 1 digit index number gets matched to the one stored in the cache but with 2 digits index number (10th class in the previous run will be matched to 9th in this run)
15:33:05 Test 60-a: Make sure that lambda classes work and get stored in the cache when another function with lambda classes in another file is being called from the current file
15:33:05 Test 60-b: Make sure that the classes are used from the cache when the program is being run again
15:33:05 -----------------------------
15:33:05
15:33:05 -----------------------------------
15:33:05 testSCCMLTests1_openj9_1_FAILED
```
[50x internal grinder](https://openj9-jenkins.osuosl.org/job/Grinder/2261)
Also seen https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_x86-64_linux_Personal/30/
FYI @hangshao0 | 1.0 | JDK21 testSCCMLTests1_openj9_1_FAILED - lambda classes - Failure link
------------
From [an internal grinder](https://openj9-jenkins.osuosl.org/job/Grinder/2261):
```
15:19:40 Eclipse OpenJ9 VM (build HEAD-1f25b601986, JRE 21 Linux amd64-64-Bit Compressed References 20230322_60 (JIT enabled, AOT enabled)
15:19:40 OpenJ9 - 1f25b601986
15:19:40 OMR - c766c4155b0
15:19:40 JCL - 732652e2726 based on jdk-21+13)
```
[Rerun in Grinder](https://openj9-jenkins.osuosl.org/job/Grinder/parambuild/?SDK_RESOURCE=customized&DOCKER_REQUIRED=false&ACTIVE_NODE_TIMEOUT=&EXTRA_DOCKER_ARGS=&PLATFORM=x86-64_linux&PERSONAL_BUILD=false&DOCKER_REGISTRY_DIR=&ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&SETUP_JCK_RUN=false&DOCKER_REGISTRY_URL_CREDENTIAL_ID=&EXTRA_OPTIONS=&BUILD_IDENTIFIER=&JENKINS_KEY=&LIGHT_WEIGHT_CHECKOUT=false&ARTIFACTORY_SERVER=ci-eclipse-openj9&USE_JRE=false&KEEP_WORKSPACE=false&JDK_VERSION=next&DOCKER_REGISTRY_URL=&EXIT_FAILURE=false&EXIT_SUCCESS=false&JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk.git&JCK_GIT_BRANCH=master&OPENJ9_BRANCH=master&OPENJ9_SHA=origin%2Fpr%2F17201%2Fmerge&VENDOR_TEST_BRANCHES=&OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&JDK_BRANCH=openj9&LABEL_ADDITION=%21sw.os.cent.6&ARTIFACTORY_REPO=ci-openj9&UPSTREAM_TEST_JOB_NUMBER=2261&DOCKERIMAGE_TAG=&RELATED_NODES=&ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&APPLICATION_OPTIONS=&CUSTOMIZED_SDK_URL_CREDENTIAL_ID=e047f50a-26c4-4b82-8525-6da17ec68450&OPENJDK_SHA=&BUILD_LIST=functional&USE_TESTENV_PROPERTIES=false&UPSTREAM_JOB_NUMBER=&STF_OWNER_BRANCH=adoptium%3Amaster&JDK_VENDOR=&BUILDS_TO_KEEP=60&TIME_LIMIT=10&PARALLEL=None&TARGET=testSCCMLTests1_openj9_1&TEST_FLAG=&UPSTREAM_TEST_JOB_NAME=Grinder&VENDOR_TEST_DIRS=&TKG_OWNER_BRANCH=adoptium%3Amaster&OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&GENERATE_JOBS=true&KEEP_REPORTDIR=false&RERUN_ITERATIONS=0&LABEL=&CUSTOMIZED_SDK_URL=https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDKnext_x86-64_linux_Personal%2F60%2FOpenJ9-JDKnext-x86-64_linux-20230322-190845.tar.gz&DAYS_TO_KEEP=7&ADOPTOPENJDK_BRANCH=master&USER_CREDENTIALS_ID=&ITERATIONS=10&VENDOR_TEST_REPOS=&JCK_GIT_REPO=&UPSTREAM_JOB_NAME=&CLOUD_PROVIDER=&CUSTOM_TARGET=&DEBUG_IMAGES_REQUIRED=false&VENDOR_TEST_SHAS=&TEST_IMAGES_REQUIRED=true&ARTIFACTORY_ROOT_DIR=Test&JDK_IMPL=openj9&TEST_TIME=&SSH_AGENT_CREDENTIAL=&AUTO_DETECT=true&SLACK_CHANNEL=&DYNAMIC_COMPILE=true&TKG_ITERATIONS=1&ARCHIVE_TEST_RESULTS=false&NUM_MACHINES=&TRSS_URL=&ADDITIONAL_ARTIFACTS_REQUIRED=&JVM_OPTIONS=) - Change TARGET to run only the failed test targets.
Optional info
-------------
Failure output (captured from console output)
---------------------------------------------
```
15:31:37 ===============================================
15:31:37 Running test testSCCMLTests1_openj9_1 ...
15:31:37 ===============================================
15:31:37 testSCCMLTests1_openj9_1 Start Time: Tue Apr 18 15:31:36 2023 Epoch Time (ms): 1681846296459
15:31:37 variation: Mode610
15:31:37 JVM_OPTIONS: -Xcompressedrefs -Xjit -Xgcpolicy:gencon
15:32:21 Testing: Test 57-a: Make sure that lambda classes work and get stored in the shared class cache
15:32:21 Test start time: 2023/04/18 15:32:20 Eastern Standard Time
15:32:21 Running command: "/home/jenkins/workspace/Grinder_iteration_0/openjdkbinary/j2sdk-image/bin/java" -Xcompressedrefs -Xjit -Xgcpolicy:gencon -Xshareclasses:name=ShareClassesCMLTests -Xtrace:print={j9shr.2259} -cp /home/jenkins/workspace/Grinder_iteration_0/aqa-tests/TKG/../../jvmtest/functional/cmdLineTests/utils/utils.jar org.openj9.test.lambdatests.Test1 0
15:32:21 Time spent starting: 3 milliseconds
15:32:22 Time spent executing: 912 milliseconds
15:32:22 Test result: FAILED
15:32:23 [OUT] 364
15:32:23 [OUT] 1720
15:32:23 [OUT] Lambda test done!
15:32:23 [ERR] 19:32:21.374*0x17000 j9shr.2259 > API j9shr_classStoreTransaction_createSharedClass : enter (classname=com/ibm/sharedclasses/spi/SharedClassProvider romClassSizeFullSize=1416 romClassMinimalSize=1416 lineNumberTableSize=0 localVariableTableSize=0
15:32:23 [ERR] 19:32:21.860 0x17000 j9shr.2259 > API j9shr_classStoreTransaction_createSharedClass : enter (classname=org/openj9/test/lambdatests/Test1$$Lambda/0x0000000000000000 romClassSizeFullSize=592 romClassMinimalSize=592 lineNumberTableSize=0 localVariableTableSize=0
15:32:23 >> Success condition was not found: [Output match: j9shr[\.]2259 > API j9shr_classStoreTransaction_createSharedClass : enter [\(]classname=org/openj9/test/lambdatests/Test1..Lambda.([\d]+)/(0x00000000|0x0000000000000000)]
15:32:23 >> Required condition was found: [Output match: Lambda test done!]
15:32:23 >> Required condition was not found: [Output match: j9shr[\.]2259 > API j9shr_classStoreTransaction_createSharedClass : enter [\(]classname=org/openj9/test/lambdatests/Test1..Lambda.([\d]+)/(0x00000000|0x0000000000000000)]
15:32:23 >> Failure condition was not found: [Output match: Unhandled Exception]
15:32:23 >> Failure condition was not found: [Output match: corrupt]
15:32:23 >> Failure condition was not found: [Output match: Processing dump event]
15:33:05 ---TEST RESULTS---
15:33:05 Number of PASSED tests: 79 out of 87
15:33:05 Number of FAILED tests: 8 out of 87
15:33:05
15:33:05 ---SUMMARY OF FAILED TESTS---
15:33:05 Test 57-a: Make sure that lambda classes work and get stored in the shared class cache
15:33:05 Test 57-b: Make sure that lambda classes are stored in the cache as orphans
15:33:05 Test 57-c: Make sure that when the program runs again lambda classes are used from the cache and not stored again
15:33:05 Test 58: Make sure that when the program runs again without the first lambda class, the second lambda class is still used from the cache and not stored again
15:33:05 Test 59-a: Run and store 10 lambda classes in the cache
15:33:05 Test 59-b: Do not run the first lambda class to check if a class with 1 digit index number gets matched to the one stored in the cache but with 2 digits index number (10th class in the previous run will be matched to 9th in this run)
15:33:05 Test 60-a: Make sure that lambda classes work and get stored in the cache when another function with lambda classes in another file is being called from the current file
15:33:05 Test 60-b: Make sure that the classes are used from the cache when the program is being run again
15:33:05 -----------------------------
15:33:05
15:33:05 -----------------------------------
15:33:05 testSCCMLTests1_openj9_1_FAILED
```
[50x internal grinder](https://openj9-jenkins.osuosl.org/job/Grinder/2261)
Also seen https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_x86-64_linux_Personal/30/
FYI @hangshao0 | non_priority | failed lambda classes failure link from eclipse vm build head jre linux bit compressed references jit enabled aot enabled omr jcl based on jdk change target to run only the failed test targets optional info failure output captured from console output running test start time tue apr epoch time ms variation jvm options xcompressedrefs xjit xgcpolicy gencon testing test a make sure that lambda classes work and get stored in the shared class cache test start time eastern standard time running command home jenkins workspace grinder iteration openjdkbinary image bin java xcompressedrefs xjit xgcpolicy gencon xshareclasses name shareclassescmltests xtrace print cp home jenkins workspace grinder iteration aqa tests tkg jvmtest functional cmdlinetests utils utils jar org test lambdatests time spent starting milliseconds time spent executing milliseconds test result failed lambda test done api classstoretransaction createsharedclass enter classname com ibm sharedclasses spi sharedclassprovider romclasssizefullsize romclassminimalsize linenumbertablesize localvariabletablesize api classstoretransaction createsharedclass enter classname org test lambdatests lambda romclasssizefullsize romclassminimalsize linenumbertablesize localvariabletablesize success condition was not found api classstoretransaction createsharedclass enter classname org test lambdatests lambda required condition was found required condition was not found api classstoretransaction createsharedclass enter classname org test lambdatests lambda failure condition was not found failure condition was not found failure condition was not found test results number of passed tests out of number of failed tests out of summary of failed tests test a make sure that lambda classes work and get stored in the shared class cache test b make sure that lambda classes are stored in the cache as orphans test c make sure that when the program runs again lambda classes are used from the cache and not stored again test make sure that when the program runs again without the first lambda class the second lambda class is still used from the cache and not stored again test a run and store lambda classes in the cache test b do not run the first lambda class to check if a class with digit index number gets matched to the one stored in the cache but with digits index number class in the previous run will be matched to in this run test a make sure that lambda classes work and get stored in the cache when another function with lambda classes in another file is being called from the current file test b make sure that the classes are used from the cache when the program is being run again failed also seen fyi | 0 |
389,397 | 26,811,389,280 | IssuesEvent | 2023-02-01 22:45:05 | RalphHightower/RalphHightower | https://api.github.com/repos/RalphHightower/RalphHightower | closed | HardwareLinks: add articles | documentation | **What page should this be added to?**<br>
HardwareLinks.md
**What section/heading should this be added to?**<br>
Articles
**Include the Markdown text that is to be added below:**<br>
HardwareLinks.md/Articles
|[Theengs open-source tools to decode BLE sensors work on ESP32, Raspberry Pi, Android phone, etc…](https://www.cnx-software.com/2022/07/22/theengs-open-source-tools-to-decode-ble-sensors-work-on-esp32-raspberry-pi-android-phone/) |
| [The All In One Review of NVIDIA Jetson Boards](https://www.iotworlds.com/the-all-in-one-review-of-nvidia-jetson-boards/) |
| [Radxa's Rock 5A Borrows the Raspberry Pi 4 Layout for an Eight-Core Beast with Up to 16GB of RAM](https://www.hackster.io/news/radxa-s-rock-5a-borrows-the-raspberry-pi-4-layout-for-an-eight-core-beast-with-up-to-16gb-of-ram-d97b583a8177)<br>$99 for the 4GB model, $119 for the 8GB model, and $159 for the 16GB model |
|[ON XGO 2 – A RASPBERRY PI CM4 BASED ROBOT DOG WITH AN ARM (CROWDFUNDING)](https://www.cnx-software.com/2023/02/01/xgo-2-a-raspberry-pi-cm4-robot-dog-robotic-arm/) |
| [WinLink E850-96Board SBC is powered by Samsung Exynos 850 Octa-core Cortex-A55 SoC](https://www.cnx-software.com/2023/02/01/winlink-e850-96board-sbc-samsung-exynos-850/?amp=1) |
**Describe alternatives you've considered**<br>
Bookmarks in browsers are not portable.
**Additional context**<br>
Add any other context or screenshots about the feature request here.
| 1.0 | HardwareLinks: add articles - **What page should this be added to?**<br>
HardwareLinks.md
**What section/heading should this be added to?**<br>
Articles
**Include the Markdown text that is to be added below:**<br>
HardwareLinks.md/Articles
|[Theengs open-source tools to decode BLE sensors work on ESP32, Raspberry Pi, Android phone, etc…](https://www.cnx-software.com/2022/07/22/theengs-open-source-tools-to-decode-ble-sensors-work-on-esp32-raspberry-pi-android-phone/) |
| [The All In One Review of NVIDIA Jetson Boards](https://www.iotworlds.com/the-all-in-one-review-of-nvidia-jetson-boards/) |
| [Radxa's Rock 5A Borrows the Raspberry Pi 4 Layout for an Eight-Core Beast with Up to 16GB of RAM](https://www.hackster.io/news/radxa-s-rock-5a-borrows-the-raspberry-pi-4-layout-for-an-eight-core-beast-with-up-to-16gb-of-ram-d97b583a8177)<br>$99 for the 4GB model, $119 for the 8GB model, and $159 for the 16GB model |
|[ON XGO 2 – A RASPBERRY PI CM4 BASED ROBOT DOG WITH AN ARM (CROWDFUNDING)](https://www.cnx-software.com/2023/02/01/xgo-2-a-raspberry-pi-cm4-robot-dog-robotic-arm/) |
| [WinLink E850-96Board SBC is powered by Samsung Exynos 850 Octa-core Cortex-A55 SoC](https://www.cnx-software.com/2023/02/01/winlink-e850-96board-sbc-samsung-exynos-850/?amp=1) |
**Describe alternatives you've considered**<br>
Bookmarks in browsers are not portable.
**Additional context**<br>
Add any other context or screenshots about the feature request here.
| non_priority | hardwarelinks add articles what page should this be added to hardwarelinks md what section heading should this be added to articles include the markdown text that is to be added below hardwarelinks md articles for the model for the model and for the model describe alternatives you ve considered bookmarks in browsers are not portable additional context add any other context or screenshots about the feature request here | 0 |
111,262 | 17,020,850,774 | IssuesEvent | 2021-07-02 18:48:14 | turkdevops/sourcegraph | https://api.github.com/repos/turkdevops/sourcegraph | opened | WS-2018-0628 (Medium) detected in marked-0.3.19.tgz | security vulnerability | ## WS-2018-0628 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.3.19.tgz">https://registry.npmjs.org/marked/-/marked-0.3.19.tgz</a></p>
<p>Path to dependency file: sourcegraph/lsif/package.json</p>
<p>Path to vulnerable library: sourcegraph/lsif/node_modules/marked</p>
<p>
Dependency Hierarchy:
- lightstep-tracer-0.24.3.tgz (Root Library)
- thrift-0.13.0.tgz
- jsdoc-3.5.5.tgz
- :x: **marked-0.3.19.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/sourcegraph/commit/e8e8af042d128ed6c81ef96fe641a69ac584771b">e8e8af042d128ed6c81ef96fe641a69ac584771b</a></p>
<p>Found in base branch: <b>dev/seed-tool</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 0.4.0 is vulnerable to Regular Expression Denial of Service (REDoS) through heading in marked.js.
<p>Publish Date: 2018-04-16
<p>URL: <a href=https://github.com/markedjs/marked/commit/09afabf69c6d0c919c03443f47bdfe476566105d>WS-2018-0628</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/markedjs/marked/releases/tag/0.4.0">https://github.com/markedjs/marked/releases/tag/0.4.0</a></p>
<p>Release Date: 2018-04-16</p>
<p>Fix Resolution: marked - 0.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2018-0628 (Medium) detected in marked-0.3.19.tgz - ## WS-2018-0628 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.3.19.tgz">https://registry.npmjs.org/marked/-/marked-0.3.19.tgz</a></p>
<p>Path to dependency file: sourcegraph/lsif/package.json</p>
<p>Path to vulnerable library: sourcegraph/lsif/node_modules/marked</p>
<p>
Dependency Hierarchy:
- lightstep-tracer-0.24.3.tgz (Root Library)
- thrift-0.13.0.tgz
- jsdoc-3.5.5.tgz
- :x: **marked-0.3.19.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/sourcegraph/commit/e8e8af042d128ed6c81ef96fe641a69ac584771b">e8e8af042d128ed6c81ef96fe641a69ac584771b</a></p>
<p>Found in base branch: <b>dev/seed-tool</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 0.4.0 is vulnerable to Regular Expression Denial of Service (REDoS) through heading in marked.js.
<p>Publish Date: 2018-04-16
<p>URL: <a href=https://github.com/markedjs/marked/commit/09afabf69c6d0c919c03443f47bdfe476566105d>WS-2018-0628</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/markedjs/marked/releases/tag/0.4.0">https://github.com/markedjs/marked/releases/tag/0.4.0</a></p>
<p>Release Date: 2018-04-16</p>
<p>Fix Resolution: marked - 0.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in marked tgz ws medium severity vulnerability vulnerable library marked tgz a markdown parser built for speed library home page a href path to dependency file sourcegraph lsif package json path to vulnerable library sourcegraph lsif node modules marked dependency hierarchy lightstep tracer tgz root library thrift tgz jsdoc tgz x marked tgz vulnerable library found in head commit a href found in base branch dev seed tool vulnerability details marked before is vulnerable to regular expression denial of service redos through heading in marked js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution marked step up your open source security game with whitesource | 0 |
245,817 | 26,569,462,604 | IssuesEvent | 2023-01-21 01:05:58 | nidhi7598/linux-3.0.35_CVE-2022-45934 | https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2022-45934 | opened | CVE-2020-35499 (Medium) detected in linux-stable-rtv3.8.6 | security vulnerability | ## CVE-2020-35499 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/sco.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A NULL pointer dereference flaw in Linux kernel versions prior to 5.11 may be seen if sco_sock_getsockopt function in net/bluetooth/sco.c do not have a sanity check for a socket connection, when using BT_SNDMTU/BT_RCVMTU for SCO sockets. This could allow a local attacker with a special user privilege to crash the system (DOS) or leak kernel internal information.
Mend Note: After conducting further research, Mend has determined that all versions of Linux Kernel up to version v5.10.4 are vulnerable to CVE-2020-35499.
<p>Publish Date: 2021-02-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-35499>CVE-2020-35499</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-35499">https://www.linuxkernelcves.com/cves/CVE-2020-35499</a></p>
<p>Release Date: 2021-02-19</p>
<p>Fix Resolution: v5.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-35499 (Medium) detected in linux-stable-rtv3.8.6 - ## CVE-2020-35499 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/sco.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A NULL pointer dereference flaw in Linux kernel versions prior to 5.11 may be seen if sco_sock_getsockopt function in net/bluetooth/sco.c do not have a sanity check for a socket connection, when using BT_SNDMTU/BT_RCVMTU for SCO sockets. This could allow a local attacker with a special user privilege to crash the system (DOS) or leak kernel internal information.
Mend Note: After conducting further research, Mend has determined that all versions of Linux Kernel up to version v5.10.4 are vulnerable to CVE-2020-35499.
<p>Publish Date: 2021-02-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-35499>CVE-2020-35499</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-35499">https://www.linuxkernelcves.com/cves/CVE-2020-35499</a></p>
<p>Release Date: 2021-02-19</p>
<p>Fix Resolution: v5.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files net bluetooth sco c vulnerability details a null pointer dereference flaw in linux kernel versions prior to may be seen if sco sock getsockopt function in net bluetooth sco c do not have a sanity check for a socket connection when using bt sndmtu bt rcvmtu for sco sockets this could allow a local attacker with a special user privilege to crash the system dos or leak kernel internal information mend note after conducting further research mend has determined that all versions of linux kernel up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
415,554 | 28,041,549,090 | IssuesEvent | 2023-03-28 18:56:03 | liveview-native/liveview-client-swiftui | https://api.github.com/repos/liveview-native/liveview-client-swiftui | closed | Documentation: SwiftUI -> User Interface Element -> Shapes: Ellipse | documentation | As a PO I'd like to have documentation for the [Ellipse](https://github.com/liveviewnative/liveview-client-swiftui/issues/83) | 1.0 | Documentation: SwiftUI -> User Interface Element -> Shapes: Ellipse - As a PO I'd like to have documentation for the [Ellipse](https://github.com/liveviewnative/liveview-client-swiftui/issues/83) | non_priority | documentation swiftui user interface element shapes ellipse as a po i d like to have documentation for the | 0 |
6,137 | 3,004,162,226 | IssuesEvent | 2015-07-25 17:00:22 | RadicalFx/Radical | https://api.github.com/repos/RadicalFx/Radical | closed | Documentation: How to handle multiple Singleton Views (Main*, Shell*) and their lifecycle | Documentation | Espceially related to the following issue https://github.com/mauroservienti/radical/issues/7#issuecomment-38457100 | 1.0 | Documentation: How to handle multiple Singleton Views (Main*, Shell*) and their lifecycle - Espceially related to the following issue https://github.com/mauroservienti/radical/issues/7#issuecomment-38457100 | non_priority | documentation how to handle multiple singleton views main shell and their lifecycle espceially related to the following issue | 0 |
42,896 | 5,544,682,063 | IssuesEvent | 2017-03-22 19:45:24 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | What language proposals would benefit from CLR changes? | Area-Language Design Discussion Feature Request Language-C# Language-VB Question | If we were to plan for one language feature that requires a revision of the CLR, then we might as well do as many of them at the same time as make sense. What changes would we consider for C# 7 and VB 15 that would benefit from CLR support? See #166 for related discussion. This is a pared-down list for us to select from.
1. Virtual extension methods (see #73, #258)
2. Generic constraints about static methods or constructors (see also #129, #154).
3. Array slicing (see #120)
4. Language support for tuples (see #347) and function types (no issue yet) might benefit from CLR unification across assemblies.
5. Traits (see #60) aka structural interfaces (see #154) or mixins (no issue yet)
Those that we would likely _not_ do in this timeframe include
1. Make void a first-class type with one value; see #234 (Suggested by @ashmind) (Given how long the CLR has been around, it is probably too late for APIs to benefit, as those that would benefit most have already been written)
2. Covariance and contravariance for classes (see #171)
3. Allow |, &, and ~ operators on a type parameter with the enum constraint (see #262)
4. Intersection types (#2146, though likely not that syntax) and/or union types.
5. Support generic indexers (see #523)
6. Higher-kinded polymorphism (see #2212)
| 1.0 | What language proposals would benefit from CLR changes? - If we were to plan for one language feature that requires a revision of the CLR, then we might as well do as many of them at the same time as make sense. What changes would we consider for C# 7 and VB 15 that would benefit from CLR support? See #166 for related discussion. This is a pared-down list for us to select from.
1. Virtual extension methods (see #73, #258)
2. Generic constraints about static methods or constructors (see also #129, #154).
3. Array slicing (see #120)
4. Language support for tuples (see #347) and function types (no issue yet) might benefit from CLR unification across assemblies.
5. Traits (see #60) aka structural interfaces (see #154) or mixins (no issue yet)
Those that we would likely _not_ do in this timeframe include
1. Make void a first-class type with one value; see #234 (Suggested by @ashmind) (Given how long the CLR has been around, it is probably too late for APIs to benefit, as those that would benefit most have already been written)
2. Covariance and contravariance for classes (see #171)
3. Allow |, &, and ~ operators on a type parameter with the enum constraint (see #262)
4. Intersection types (#2146, though likely not that syntax) and/or union types.
5. Support generic indexers (see #523)
6. Higher-kinded polymorphism (see #2212)
| non_priority | what language proposals would benefit from clr changes if we were to plan for one language feature that requires a revision of the clr then we might as well do as many of them at the same time as make sense what changes would we consider for c and vb that would benefit from clr support see for related discussion this is a pared down list for us to select from virtual extension methods see generic constraints about static methods or constructors see also array slicing see language support for tuples see and function types no issue yet might benefit from clr unification across assemblies traits see aka structural interfaces see or mixins no issue yet those that we would likely not do in this timeframe include make void a first class type with one value see suggested by ashmind given how long the clr has been around it is probably too late for apis to benefit as those that would benefit most have already been written covariance and contravariance for classes see allow and operators on a type parameter with the enum constraint see intersection types though likely not that syntax and or union types support generic indexers see higher kinded polymorphism see | 0 |
339,486 | 24,620,874,335 | IssuesEvent | 2022-10-15 22:47:27 | citizen00147/atlanta-tech-resources | https://api.github.com/repos/citizen00147/atlanta-tech-resources | opened | More Content Needed | documentation enhancement good first issue | Anyone is free to contribute so long as the contents are specific to Atlanta, GA and the metro area! I know I'm missing a lot of stuff so if you know of something, please open a PR, but check out the [contributing docs](https://github.com/citizen00147/atlanta-tech-resources/blob/master/docs/CONTRIBUTING.md) first! | 1.0 | More Content Needed - Anyone is free to contribute so long as the contents are specific to Atlanta, GA and the metro area! I know I'm missing a lot of stuff so if you know of something, please open a PR, but check out the [contributing docs](https://github.com/citizen00147/atlanta-tech-resources/blob/master/docs/CONTRIBUTING.md) first! | non_priority | more content needed anyone is free to contribute so long as the contents are specific to atlanta ga and the metro area i know i m missing a lot of stuff so if you know of something please open a pr but check out the first | 0 |
255,925 | 19,346,305,115 | IssuesEvent | 2021-12-15 11:10:22 | ostra-project/ostra-vm | https://api.github.com/repos/ostra-project/ostra-vm | opened | updateFee() function -> Compromised Account Issue | documentation security-problem | Issue Source: [HashEx](https://blog.hashex.org/safemoon-smart-contract-audit-report-8e4b843a375d)
1. 'Temporary ownership renounce' -> Solved by using Diamond Proxy with a governance system.
2. 'No safeguards for fees and maxTxAmount':
```SOLIDITY
require((_liquidityFee + _retribFee + _burnFee) < 10); // Fees are limited to 10 MAX
require(amount <= XXX); // MaxTxAmount is limited to 0.5% of the total supply
require(amount <= XXX); // NumTokensSellToAddToLiquidity is limited to 0.05% of the total supply
```
3. 'excludeFromReward() abuse' -> Solved in the [Issue #3](https://github.com/ostra-project/ostra-vm/issues/3)
4. 'excluded[] length problem' -> UNSOLVED
5. 'ERC20 standard violation' -> UNSOLVED
6. 'Locked ether' -> UNSOLVED
7. 'addLiquidity() recipient' -> UNSOLVED
8. 'Hardcoded addresses' -> UNSOLVED
9. 'inSwapAndLiquify visibility' -> UNSOLVED
10. 'numTokensSellToAddToLiquidity is constant' -> UNSOLVED
11. 'Incorrect error message' -> `Account is already excluded` replaced by `Sender is already included`
12. General recommendations:
- _removeFee() & _restoreFee() are optimized (Fees replaced by a struct, no variable declared inside these functions)
- contractTokenBalance and useless condition has been removed from the _transfer() function
- Unused Address library has been removed
| 1.0 | updateFee() function -> Compromised Account Issue - Issue Source: [HashEx](https://blog.hashex.org/safemoon-smart-contract-audit-report-8e4b843a375d)
1. 'Temporary ownership renounce' -> Solved by using Diamond Proxy with a governance system.
2. 'No safeguards for fees and maxTxAmount':
```SOLIDITY
require((_liquidityFee + _retribFee + _burnFee) < 10); // Fees are limited to 10 MAX
require(amount <= XXX); // MaxTxAmount is limited to 0.5% of the total supply
require(amount <= XXX); // NumTokensSellToAddToLiquidity is limited to 0.05% of the total supply
```
3. 'excludeFromReward() abuse' -> Solved in the [Issue #3](https://github.com/ostra-project/ostra-vm/issues/3)
4. 'excluded[] length problem' -> UNSOLVED
5. 'ERC20 standard violation' -> UNSOLVED
6. 'Locked ether' -> UNSOLVED
7. 'addLiquidity() recipient' -> UNSOLVED
8. 'Hardcoded addresses' -> UNSOLVED
9. 'inSwapAndLiquify visibility' -> UNSOLVED
10. 'numTokensSellToAddToLiquidity is constant' -> UNSOLVED
11. 'Incorrect error message' -> `Account is already excluded` replaced by `Sender is already included`
12. General recommendations:
- _removeFee() & _restoreFee() are optimized (Fees replaced by a struct, no variable declared inside these functions)
- contractTokenBalance and useless condition has been removed from the _transfer() function
- Unused Address library has been removed
| non_priority | updatefee function compromised account issue issue source temporary ownership renounce solved by using diamond proxy with a governance system no safeguards for fees and maxtxamount solidity require liquidityfee retribfee burnfee fees are limited to max require amount xxx maxtxamount is limited to of the total supply require amount xxx numtokensselltoaddtoliquidity is limited to of the total supply excludefromreward abuse solved in the excluded length problem unsolved standard violation unsolved locked ether unsolved addliquidity recipient unsolved hardcoded addresses unsolved inswapandliquify visibility unsolved numtokensselltoaddtoliquidity is constant unsolved incorrect error message account is already excluded replaced by sender is already included general recommendations removefee restorefee are optimized fees replaced by a struct no variable declared inside these functions contracttokenbalance and useless condition has been removed from the transfer function unused address library has been removed | 0 |
338,553 | 24,590,114,329 | IssuesEvent | 2022-10-14 00:47:45 | IvanJosipovic/ingress-nginx-validate-jwt | https://api.github.com/repos/IvanJosipovic/ingress-nginx-validate-jwt | closed | Docker Image | documentation | By any chance would you be able to provider the Dockerfile? (perhaps place it in the root)
Thanks :) | 1.0 | Docker Image - By any chance would you be able to provider the Dockerfile? (perhaps place it in the root)
Thanks :) | non_priority | docker image by any chance would you be able to provider the dockerfile perhaps place it in the root thanks | 0 |
113,316 | 11,799,191,804 | IssuesEvent | 2020-03-18 15:31:34 | Azure-Samples/durable-functions-producer-consumer | https://api.github.com/repos/Azure-Samples/durable-functions-producer-consumer | closed | Create Kafka producer | documentation enhancement good first issue help wanted | ```
- [ ] bug report -> please search issues before submitting
- [x] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
```
It would be useful to have a producer for Kafka endpoints! | 1.0 | Create Kafka producer - ```
- [ ] bug report -> please search issues before submitting
- [x] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
```
It would be useful to have a producer for Kafka endpoints! | non_priority | create kafka producer bug report please search issues before submitting feature request documentation issue or request regression a behavior that used to work and stopped in a new release it would be useful to have a producer for kafka endpoints | 0 |
258,529 | 27,564,253,282 | IssuesEvent | 2023-03-08 01:39:05 | victorlmneves/search-team-javascript | https://api.github.com/repos/victorlmneves/search-team-javascript | opened | CVE-2017-20165 (High) detected in debug-0.8.1.tgz | security vulnerability | ## CVE-2017-20165 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-0.8.1.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-0.8.1.tgz">https://registry.npmjs.org/debug/-/debug-0.8.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/tiny-lr/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- gulp-webserver-0.9.1.tgz (Root Library)
- tiny-lr-0.1.4.tgz
- :x: **debug-0.8.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability classified as problematic has been found in debug-js debug up to 3.0.x. This affects the function useColors of the file src/node.js. The manipulation of the argument str leads to inefficient regular expression complexity. Upgrading to version 3.1.0 is able to address this issue. The name of the patch is c38a0166c266a679c8de012d4eaccec3f944e685. It is recommended to upgrade the affected component. The identifier VDB-217665 was assigned to this vulnerability.
<p>Publish Date: 2023-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-20165>CVE-2017-20165</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9vvw-cc9w-f27h">https://github.com/advisories/GHSA-9vvw-cc9w-f27h</a></p>
<p>Release Date: 2023-01-09</p>
<p>Fix Resolution: debug - 2.6.9,3.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-20165 (High) detected in debug-0.8.1.tgz - ## CVE-2017-20165 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-0.8.1.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-0.8.1.tgz">https://registry.npmjs.org/debug/-/debug-0.8.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/tiny-lr/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- gulp-webserver-0.9.1.tgz (Root Library)
- tiny-lr-0.1.4.tgz
- :x: **debug-0.8.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability classified as problematic has been found in debug-js debug up to 3.0.x. This affects the function useColors of the file src/node.js. The manipulation of the argument str leads to inefficient regular expression complexity. Upgrading to version 3.1.0 is able to address this issue. The name of the patch is c38a0166c266a679c8de012d4eaccec3f944e685. It is recommended to upgrade the affected component. The identifier VDB-217665 was assigned to this vulnerability.
<p>Publish Date: 2023-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-20165>CVE-2017-20165</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9vvw-cc9w-f27h">https://github.com/advisories/GHSA-9vvw-cc9w-f27h</a></p>
<p>Release Date: 2023-01-09</p>
<p>Fix Resolution: debug - 2.6.9,3.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in debug tgz cve high severity vulnerability vulnerable library debug tgz small debugging utility library home page a href path to dependency file package json path to vulnerable library node modules tiny lr node modules debug package json dependency hierarchy gulp webserver tgz root library tiny lr tgz x debug tgz vulnerable library found in base branch master vulnerability details a vulnerability classified as problematic has been found in debug js debug up to x this affects the function usecolors of the file src node js the manipulation of the argument str leads to inefficient regular expression complexity upgrading to version is able to address this issue the name of the patch is it is recommended to upgrade the affected component the identifier vdb was assigned to this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution debug step up your open source security game with mend | 0 |
262,450 | 27,946,597,917 | IssuesEvent | 2023-03-24 04:03:15 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [Security Solution]: Alert treemap, table and trend graph retains when user change the alert status from alert view details flyout. | bug impact:medium Team:Threat Hunting fixed Team: SecuritySolution Team:Threat Hunting:Investigations v8.8.0 | **Describe the bug:**
Alert treemap, table and trend graph retains when user change the alert status from alert view details flyout.
**Build Details:**
```
VERSION: 8.5.0-BC6
BUILD: 57022
COMMIT: 0b06c1437f38a9eb9a4273003d8d2ce2c8905189
```
**Preconditions**
1. Kibana should be running.
**Steps to Reproduce**
1. Navigate to Alerts tab.
2. Now, select an alert and click on view details.
3. Now, change the alert status on alert details flyout.
4. Observe that alert treemap, table and trend graph retains when user change the alert status from alert view details flyout.
**Actual Result**
Alert treemap, table and trend graph retains when user change the alert status from alert view details flyout.
**Expected Result**
Alert treemap, table and trend graph should not retain when user change the alert status from alert view details flyout.
**Screen-Recording:**
https://user-images.githubusercontent.com/84007952/196669519-c65e64f3-3c29-4e73-933b-f725922013ba.mp4 | True | [Security Solution]: Alert treemap, table and trend graph retains when user change the alert status from alert view details flyout. - **Describe the bug:**
Alert treemap, table and trend graph retains when user change the alert status from alert view details flyout.
**Build Details:**
```
VERSION: 8.5.0-BC6
BUILD: 57022
COMMIT: 0b06c1437f38a9eb9a4273003d8d2ce2c8905189
```
**Preconditions**
1. Kibana should be running.
**Steps to Reproduce**
1. Navigate to Alerts tab.
2. Now, select an alert and click on view details.
3. Now, change the alert status on alert details flyout.
4. Observe that alert treemap, table and trend graph retains when user change the alert status from alert view details flyout.
**Actual Result**
Alert treemap, table and trend graph retains when user change the alert status from alert view details flyout.
**Expected Result**
Alert treemap, table and trend graph should not retain when user change the alert status from alert view details flyout.
**Screen-Recording:**
https://user-images.githubusercontent.com/84007952/196669519-c65e64f3-3c29-4e73-933b-f725922013ba.mp4 | non_priority | alert treemap table and trend graph retains when user change the alert status from alert view details flyout describe the bug alert treemap table and trend graph retains when user change the alert status from alert view details flyout build details version build commit preconditions kibana should be running steps to reproduce navigate to alerts tab now select an alert and click on view details now change the alert status on alert details flyout observe that alert treemap table and trend graph retains when user change the alert status from alert view details flyout actual result alert treemap table and trend graph retains when user change the alert status from alert view details flyout expected result alert treemap table and trend graph should not retain when user change the alert status from alert view details flyout screen recording | 0 |
57,757 | 14,217,822,753 | IssuesEvent | 2020-11-17 10:53:57 | numba/llvmlite | https://api.github.com/repos/numba/llvmlite | closed | Build breaks on FreeBSD: it needs -fPIC | build needtriage | ```
got version from file /wrkdirs/usr/ports/devel/py-llvmlite/work-py36/llvmlite-0.31.0/llvmlite/_version.py {'version': '0.31.0', 'full': 'fe7d985f6421d87f613bd414479d29d912771562'}
running build_ext
/usr/local/bin/python3.6 /wrkdirs/usr/ports/devel/py-llvmlite/work-py36/llvmlite-0.31.0/ffi/build.py
LLVM version... clang++ -std=c++11 -stdlib=libc++ -shared -I/usr/local/llvm80/include -std=c++11 -fno-exceptions -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -fno-rtti -g assembly.cpp bitcode.cpp core.cpp initfini.cpp module.cpp value.cpp executionengine.cpp transforms.cpp passmanagers.cpp targets.cpp dylib.cpp linker.cpp object_file.cpp -o libllvmlite.so -L/usr/local/llvm80/lib -lLLVM-8
ld: error: can't create dynamic relocation R_X86_64_64 against local symbol in readonly segment; recompile object files with -fPIC or pass '-Wl,-z,notext' to allow text relocations in the output
```
using cmake to build
llvmlite-0.31.0 | 1.0 | Build breaks on FreeBSD: it needs -fPIC - ```
got version from file /wrkdirs/usr/ports/devel/py-llvmlite/work-py36/llvmlite-0.31.0/llvmlite/_version.py {'version': '0.31.0', 'full': 'fe7d985f6421d87f613bd414479d29d912771562'}
running build_ext
/usr/local/bin/python3.6 /wrkdirs/usr/ports/devel/py-llvmlite/work-py36/llvmlite-0.31.0/ffi/build.py
LLVM version... clang++ -std=c++11 -stdlib=libc++ -shared -I/usr/local/llvm80/include -std=c++11 -fno-exceptions -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -fno-rtti -g assembly.cpp bitcode.cpp core.cpp initfini.cpp module.cpp value.cpp executionengine.cpp transforms.cpp passmanagers.cpp targets.cpp dylib.cpp linker.cpp object_file.cpp -o libllvmlite.so -L/usr/local/llvm80/lib -lLLVM-8
ld: error: can't create dynamic relocation R_X86_64_64 against local symbol in readonly segment; recompile object files with -fPIC or pass '-Wl,-z,notext' to allow text relocations in the output
```
using cmake to build
llvmlite-0.31.0 | non_priority | build breaks on freebsd it needs fpic got version from file wrkdirs usr ports devel py llvmlite work llvmlite llvmlite version py version full running build ext usr local bin wrkdirs usr ports devel py llvmlite work llvmlite ffi build py llvm version clang std c stdlib libc shared i usr local include std c fno exceptions d stdc constant macros d stdc format macros d stdc limit macros fno rtti g assembly cpp bitcode cpp core cpp initfini cpp module cpp value cpp executionengine cpp transforms cpp passmanagers cpp targets cpp dylib cpp linker cpp object file cpp o libllvmlite so l usr local lib lllvm ld error can t create dynamic relocation r against local symbol in readonly segment recompile object files with fpic or pass wl z notext to allow text relocations in the output using cmake to build llvmlite | 0 |
306,295 | 26,455,354,817 | IssuesEvent | 2023-01-16 14:06:46 | GeorgievaSilvia/Help-a-Paw | https://api.github.com/repos/GeorgievaSilvia/Help-a-Paw | opened | Emergency signal submitting from different location | positive test case | Pre-conditions:
1.The user must be connected to the Internet.
2.The user is logged in to the application.
Steps:
1
Tap to different place on the map
Expected result
Report window is opened successfully
2
Fill description field
Expected result
Description field is filled
3
Fill phone number
Expected result
Pnone number is filled
4
Tap dropdown menu
Expected result
Dropdown menu with available signals is shown on the screen
5
Select "Emergency signal"
Expected result
"Emergency signal" is marked
6
Tap "Send" button
Expected result
The signal is successfully submitted | 1.0 | Emergency signal submitting from different location - Pre-conditions:
1.The user must be connected to the Internet.
2.The user is logged in to the application.
Steps:
1
Tap to different place on the map
Expected result
Report window is opened successfully
2
Fill description field
Expected result
Description field is filled
3
Fill phone number
Expected result
Pnone number is filled
4
Tap dropdown menu
Expected result
Dropdown menu with available signals is shown on the screen
5
Select "Emergency signal"
Expected result
"Emergency signal" is marked
6
Tap "Send" button
Expected result
The signal is successfully submitted | non_priority | emergency signal submitting from different location pre conditions the user must be connected to the internet the user is logged in to the application steps tap to different place on the map expected result report window is opened successfully fill description field expected result description field is filled fill phone number expected result pnone number is filled tap dropdown menu expected result dropdown menu with available signals is shown on the screen select emergency signal expected result emergency signal is marked tap send button expected result the signal is successfully submitted | 0 |
31,743 | 8,742,620,783 | IssuesEvent | 2018-12-12 16:51:34 | MTG/essentia | https://api.github.com/repos/MTG/essentia | closed | ./waf build fail - TagLib | builds | Hello,
I'm trying to run the script ./waf and when I use flags `--mode=release --build-static --with-python --with-cpptests --with-examples --with-vamp`, I always get stuck at the file metadatareader.cpp. Stacktrace:
`[338/374] Linking build/src/examples/essentia_standard_beatsmarker
[339/374] Linking build/src/examples/essentia_standard_onsetrate
src/libessentia.a(metadatareader.cpp.1.o): In function `formatString(TagLib::StringList const&)':
metadatareader.cpp:(.text+0x14f1): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x157d): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x15b4): undefined reference to `TagLib::String::to8Bit(bool) const'
src/libessentia.a(metadatareader.cpp.1.o): In function `essentia::standard::MetadataReader::compute()':
metadatareader.cpp:(.text+0x2a85): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x2c67): undefined reference to `TagLib::String::to8Bit(bool) const'
collect2: error: ld returned 1 exit status
src/libessentia.a(metadatareader.cpp.1.o): In function `formatString(TagLib::StringList const&)':
metadatareader.cpp:(.text+0x14f1): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x157d): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x15b4): undefined reference to `TagLib::String::to8Bit(bool) const'
src/libessentia.a(metadatareader.cpp.1.o): In function `essentia::standard::MetadataReader::compute()':
metadatareader.cpp:(.text+0x2a85): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x2c67): undefined reference to `TagLib::String::to8Bit(bool) const'
collect2: error: ld returned 1 exit status
Waf: Leaving directory `/home/kapi/essentia/build'
Build failed
-> task in 'essentia_standard_beatsmarker' failed with exit status 1 (run with -v to display more information)
-> task in 'essentia_standard_onsetrate' failed with exit status 1 (run with -v to display more information)
`
I tried installing both the newest (1.11.1) and one of the older (1.9) versions of the TagLib. What can I do to make it work? My operating system is Ubuntu 16.04 LTS. | 1.0 | ./waf build fail - TagLib - Hello,
I'm trying to run the script ./waf and when I use flags `--mode=release --build-static --with-python --with-cpptests --with-examples --with-vamp`, I always get stuck at the file metadatareader.cpp. Stacktrace:
`[338/374] Linking build/src/examples/essentia_standard_beatsmarker
[339/374] Linking build/src/examples/essentia_standard_onsetrate
src/libessentia.a(metadatareader.cpp.1.o): In function `formatString(TagLib::StringList const&)':
metadatareader.cpp:(.text+0x14f1): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x157d): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x15b4): undefined reference to `TagLib::String::to8Bit(bool) const'
src/libessentia.a(metadatareader.cpp.1.o): In function `essentia::standard::MetadataReader::compute()':
metadatareader.cpp:(.text+0x2a85): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x2c67): undefined reference to `TagLib::String::to8Bit(bool) const'
collect2: error: ld returned 1 exit status
src/libessentia.a(metadatareader.cpp.1.o): In function `formatString(TagLib::StringList const&)':
metadatareader.cpp:(.text+0x14f1): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x157d): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x15b4): undefined reference to `TagLib::String::to8Bit(bool) const'
src/libessentia.a(metadatareader.cpp.1.o): In function `essentia::standard::MetadataReader::compute()':
metadatareader.cpp:(.text+0x2a85): undefined reference to `TagLib::String::to8Bit(bool) const'
metadatareader.cpp:(.text+0x2c67): undefined reference to `TagLib::String::to8Bit(bool) const'
collect2: error: ld returned 1 exit status
Waf: Leaving directory `/home/kapi/essentia/build'
Build failed
-> task in 'essentia_standard_beatsmarker' failed with exit status 1 (run with -v to display more information)
-> task in 'essentia_standard_onsetrate' failed with exit status 1 (run with -v to display more information)
`
I tried installing both the newest (1.11.1) and one of the older (1.9) versions of the TagLib. What can I do to make it work? My operating system is Ubuntu 16.04 LTS. | non_priority | waf build fail taglib hello i m trying to run the script waf and when i use flags mode release build static with python with cpptests with examples with vamp i always get stuck at the file metadatareader cpp stacktrace linking build src examples essentia standard beatsmarker linking build src examples essentia standard onsetrate src libessentia a metadatareader cpp o in function formatstring taglib stringlist const metadatareader cpp text undefined reference to taglib string bool const metadatareader cpp text undefined reference to taglib string bool const metadatareader cpp text undefined reference to taglib string bool const src libessentia a metadatareader cpp o in function essentia standard metadatareader compute metadatareader cpp text undefined reference to taglib string bool const metadatareader cpp text undefined reference to taglib string bool const error ld returned exit status src libessentia a metadatareader cpp o in function formatstring taglib stringlist const metadatareader cpp text undefined reference to taglib string bool const metadatareader cpp text undefined reference to taglib string bool const metadatareader cpp text undefined reference to taglib string bool const src libessentia a metadatareader cpp o in function essentia standard metadatareader compute metadatareader cpp text undefined reference to taglib string bool const metadatareader cpp text undefined reference to taglib string bool const error ld returned exit status waf leaving directory home kapi essentia build build failed task in essentia standard beatsmarker failed with exit status run with v to display more information task in essentia standard onsetrate failed with exit status run with v to display more information i tried installing both the newest and one of the older versions of the taglib what can i do to make it work my operating system is ubuntu lts | 0 |
85,244 | 24,545,966,783 | IssuesEvent | 2022-10-12 08:49:20 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [VAPT] [SB] Create / Edit study > Study Information > Study type field is getting displayed in the study information screen | Bug P1 Study builder Process: Fixed Process: Tested QA Process: Tested dev | **AR:** Create / Edit study > Study Information > Study type field is getting displayed in the study information screen
**ER:** Create / Edit study > Study Information > Study type field should not get displayed in the study information screen

| 1.0 | [VAPT] [SB] Create / Edit study > Study Information > Study type field is getting displayed in the study information screen - **AR:** Create / Edit study > Study Information > Study type field is getting displayed in the study information screen
**ER:** Create / Edit study > Study Information > Study type field should not get displayed in the study information screen

| non_priority | create edit study study information study type field is getting displayed in the study information screen ar create edit study study information study type field is getting displayed in the study information screen er create edit study study information study type field should not get displayed in the study information screen | 0 |
159,705 | 13,770,480,202 | IssuesEvent | 2020-10-07 20:18:19 | sudojunior/journey | https://api.github.com/repos/sudojunior/journey | opened | rfc: Enhanced implementation for entry metadata | discussion documentation enhancement rfc | Some sites may not have developer *friendly* interfaces. As an alternative, a link can be used in its place.
- Requiring a script or extra context to be added on top of what has already been imported (i.e. Scripts, Footnote, etc.)
- Requiring to interface with an API and build the embed based on the data response.
## Media Deployment Platforms
- Dezzer
- Last.fm (/:Artist+Name/_/:Record+Name)
- Soundcloud (/:artist/:title)
- Spotify (*multiple formats*)
https://developer.spotify.com/documentation/widgets/
- Artist (includes follow button)
- Album (Large at w:300px,h:380px & Compact at w:300px,h:80px)
- [x] Playlist
- Podcast
- [x] Track
- [x] YouTube (unified content display with multiple platforms)
- Music
- Gaming
- Twitch
- Stream
- [x] Clip
- [x] VOD
- Twitter
- Tweet
- User (+ follow button)
## Others to note
- https://badgen.net/
- https://shields.io/ | 1.0 | rfc: Enhanced implementation for entry metadata - Some sites may not have developer *friendly* interfaces. As an alternative, a link can be used in its place.
- Requiring a script or extra context to be added on top of what has already been imported (i.e. Scripts, Footnote, etc.)
- Requiring to interface with an API and build the embed based on the data response.
## Media Deployment Platforms
- Dezzer
- Last.fm (/:Artist+Name/_/:Record+Name)
- Soundcloud (/:artist/:title)
- Spotify (*multiple formats*)
https://developer.spotify.com/documentation/widgets/
- Artist (includes follow button)
- Album (Large at w:300px,h:380px & Compact at w:300px,h:80px)
- [x] Playlist
- Podcast
- [x] Track
- [x] YouTube (unified content display with multiple platforms)
- Music
- Gaming
- Twitch
- Stream
- [x] Clip
- [x] VOD
- Twitter
- Tweet
- User (+ follow button)
## Others to note
- https://badgen.net/
- https://shields.io/ | non_priority | rfc enhanced implementation for entry metadata some sites may not have developer friendly interfaces as an alternative a link can be used in its place requiring a script or extra context to be added on top of what has already been imported i e scripts footnote etc requiring to interface with an api and build the embed based on the data response media deployment platforms dezzer last fm artist name record name soundcloud artist title spotify multiple formats artist includes follow button album large at w h compact at w h playlist podcast track youtube unified content display with multiple platforms music gaming twitch stream clip vod twitter tweet user follow button others to note | 0 |
91,355 | 10,719,376,657 | IssuesEvent | 2019-10-26 09:21:50 | Timotheeee/PSIT3-HS19-IT18a_ZH-TeamIP2 | https://api.github.com/repos/Timotheeee/PSIT3-HS19-IT18a_ZH-TeamIP2 | closed | Document intial payload | documentation | Define the payload that index.html receives upon first rendering.
This should include all necessary init JS code as well as **all** data (JSON) necessary for it to function.
-> Also document the requests, which will happen.
| 1.0 | Document intial payload - Define the payload that index.html receives upon first rendering.
This should include all necessary init JS code as well as **all** data (JSON) necessary for it to function.
-> Also document the requests, which will happen.
| non_priority | document intial payload define the payload that index html receives upon first rendering this should include all necessary init js code as well as all data json necessary for it to function also document the requests which will happen | 0 |
90,872 | 8,285,832,757 | IssuesEvent | 2018-09-19 01:13:55 | ccadllc/cedi-dtrace | https://api.github.com/repos/ccadllc/cedi-dtrace | closed | Add tests validating the laws of the cats/cats-effect typeclass instances here | Tests | Provide tests to validate the TC instance laws, using scalacheck and discipline. | 1.0 | Add tests validating the laws of the cats/cats-effect typeclass instances here - Provide tests to validate the TC instance laws, using scalacheck and discipline. | non_priority | add tests validating the laws of the cats cats effect typeclass instances here provide tests to validate the tc instance laws using scalacheck and discipline | 0 |
53,682 | 23,031,383,943 | IssuesEvent | 2022-07-22 14:14:06 | nftstorage/nft.storage | https://api.github.com/repos/nftstorage/nft.storage | closed | Continuous Pinata backup job on Heroku | kind/enhancement P3 stack/write-services | Issue for similar work to: https://github.com/web3-storage/web3.storage/pull/411
Basically trying to pin to Pinata continuously so we fall less behind given its rate limits. | 1.0 | Continuous Pinata backup job on Heroku - Issue for similar work to: https://github.com/web3-storage/web3.storage/pull/411
Basically trying to pin to Pinata continuously so we fall less behind given its rate limits. | non_priority | continuous pinata backup job on heroku issue for similar work to basically trying to pin to pinata continuously so we fall less behind given its rate limits | 0 |
189,472 | 14,504,438,246 | IssuesEvent | 2020-12-12 01:00:50 | kalexmills/github-vet-tests-dec2020 | https://api.github.com/repos/kalexmills/github-vet-tests-dec2020 | closed | presslabs/mysql-operator: pkg/controller/mysqlcluster/internal/syncer/pod_test.go; 3 LoC | fresh test tiny |
Found a possible issue in [presslabs/mysql-operator](https://www.github.com/presslabs/mysql-operator) at [pkg/controller/mysqlcluster/internal/syncer/pod_test.go](https://github.com/presslabs/mysql-operator/blob/47436cc5edd3d974239e45c994ccf7e22afa65e0/pkg/controller/mysqlcluster/internal/syncer/pod_test.go#L86-L88)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to pod at line 87 may start a goroutine
[Click here to see the code in its original context.](https://github.com/presslabs/mysql-operator/blob/47436cc5edd3d974239e45c994ccf7e22afa65e0/pkg/controller/mysqlcluster/internal/syncer/pod_test.go#L86-L88)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, pod := range podList.Items {
c.Delete(context.TODO(), &pod)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 47436cc5edd3d974239e45c994ccf7e22afa65e0
| 1.0 | presslabs/mysql-operator: pkg/controller/mysqlcluster/internal/syncer/pod_test.go; 3 LoC -
Found a possible issue in [presslabs/mysql-operator](https://www.github.com/presslabs/mysql-operator) at [pkg/controller/mysqlcluster/internal/syncer/pod_test.go](https://github.com/presslabs/mysql-operator/blob/47436cc5edd3d974239e45c994ccf7e22afa65e0/pkg/controller/mysqlcluster/internal/syncer/pod_test.go#L86-L88)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to pod at line 87 may start a goroutine
[Click here to see the code in its original context.](https://github.com/presslabs/mysql-operator/blob/47436cc5edd3d974239e45c994ccf7e22afa65e0/pkg/controller/mysqlcluster/internal/syncer/pod_test.go#L86-L88)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, pod := range podList.Items {
c.Delete(context.TODO(), &pod)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 47436cc5edd3d974239e45c994ccf7e22afa65e0
| non_priority | presslabs mysql operator pkg controller mysqlcluster internal syncer pod test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to pod at line may start a goroutine click here to show the line s of go which triggered the analyzer go for pod range podlist items c delete context todo pod leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
171,997 | 21,015,904,817 | IssuesEvent | 2022-03-30 10:58:58 | ministryofjustice/cloud-operations | https://api.github.com/repos/ministryofjustice/cloud-operations | opened | 🕵️ ECS services should not have public IP addresses assigned to them automatically | security | Checks if ECS services are configured to automatically assign public IP addresses. This check fails if AssignPublicIP is ENABLED.
Source
[AWS Security Hub](https://console.aws.amazon.com/securityhub/home)
Security Hub control ID: ECS.2
Alert Criteria
Red: Critical or High. Security Hub control failed.
### Definition of Done
Establish which accounts this applies to and resolve in each
- [ ] Development
- [ ] Pre-production
- [ ] Production
Recommended Action
Follow the [Security Hub documentation](https://docs.aws.amazon.com/console/securityhub/ECS.2/remediation) to fix the issue.
| True | 🕵️ ECS services should not have public IP addresses assigned to them automatically - Checks if ECS services are configured to automatically assign public IP addresses. This check fails if AssignPublicIP is ENABLED.
Source
[AWS Security Hub](https://console.aws.amazon.com/securityhub/home)
Security Hub control ID: ECS.2
Alert Criteria
Red: Critical or High. Security Hub control failed.
### Definition of Done
Establish which accounts this applies to and resolve in each
- [ ] Development
- [ ] Pre-production
- [ ] Production
Recommended Action
Follow the [Security Hub documentation](https://docs.aws.amazon.com/console/securityhub/ECS.2/remediation) to fix the issue.
| non_priority | 🕵️ ecs services should not have public ip addresses assigned to them automatically checks if ecs services are configured to automatically assign public ip addresses this check fails if assignpublicip is enabled source security hub control id ecs alert criteria red critical or high security hub control failed definition of done establish which accounts this applies to and resolve in each development pre production production recommended action follow the to fix the issue | 0 |
345,062 | 24,842,129,371 | IssuesEvent | 2022-10-26 13:28:26 | pactumjs/pactumjs.github.io | https://api.github.com/repos/pactumjs/pactumjs.github.io | closed | Add ENV page | documentation | Create a new page that will list all the environment variables supported by pactum with their usage and example.
A single page is sufficient.
https://github.com/pactumjs/pactum/blob/master/src/config.js | 1.0 | Add ENV page - Create a new page that will list all the environment variables supported by pactum with their usage and example.
A single page is sufficient.
https://github.com/pactumjs/pactum/blob/master/src/config.js | non_priority | add env page create a new page that will list all the environment variables supported by pactum with their usage and example a single page is sufficient | 0 |
40,089 | 16,327,396,615 | IssuesEvent | 2021-05-12 03:58:58 | MicrosoftDocs/dynamics-365-customer-engagement | https://api.github.com/repos/MicrosoftDocs/dynamics-365-customer-engagement | closed | Customizing the Timeline control | Pri1 assigned-to-author dynamics-365-customerservice/svc | Hello team!!!
I think it would be great if can select which fields are available to filter timeline records. For instance, today I had a request to add the Owner field and it seems that is not possible as it is explained [here](https://docs.microsoft.com/en-us/dynamics365/customer-service/customer-service-hub-user-guide-timeline-admin#filter-records-on-timeline).
Would you consider this as an improvement? Or is this possible and I am missing something?
Thanks,
Ale
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 680e01fe-96e2-f539-be10-c8518942d090
* Version Independent ID: 44a23653-1f66-f638-ec5c-498f4b877bfb
* Content: [Configure Timeline](https://docs.microsoft.com/en-us/dynamics365/customer-service/customer-service-hub-user-guide-timeline-admin)
* Content Source: [ce/customer-service/customer-service-hub-user-guide-timeline-admin.md](https://github.com/MicrosoftDocs/dynamics-365-customer-engagement/blob/main/ce/customer-service/customer-service-hub-user-guide-timeline-admin.md)
* Service: **dynamics-365-customerservice**
* GitHub Login: @lalexms
* Microsoft Alias: **laalexan** | 1.0 | Customizing the Timeline control - Hello team!!!
I think it would be great if can select which fields are available to filter timeline records. For instance, today I had a request to add the Owner field and it seems that is not possible as it is explained [here](https://docs.microsoft.com/en-us/dynamics365/customer-service/customer-service-hub-user-guide-timeline-admin#filter-records-on-timeline).
Would you consider this as an improvement? Or is this possible and I am missing something?
Thanks,
Ale
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 680e01fe-96e2-f539-be10-c8518942d090
* Version Independent ID: 44a23653-1f66-f638-ec5c-498f4b877bfb
* Content: [Configure Timeline](https://docs.microsoft.com/en-us/dynamics365/customer-service/customer-service-hub-user-guide-timeline-admin)
* Content Source: [ce/customer-service/customer-service-hub-user-guide-timeline-admin.md](https://github.com/MicrosoftDocs/dynamics-365-customer-engagement/blob/main/ce/customer-service/customer-service-hub-user-guide-timeline-admin.md)
* Service: **dynamics-365-customerservice**
* GitHub Login: @lalexms
* Microsoft Alias: **laalexan** | non_priority | customizing the timeline control hello team i think it would be great if can select which fields are available to filter timeline records for instance today i had a request to add the owner field and it seems that is not possible as it is explained would you consider this as an improvement or is this possible and i am missing something thanks ale document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service dynamics customerservice github login lalexms microsoft alias laalexan | 0 |
75,391 | 7,470,262,518 | IssuesEvent | 2018-04-03 03:49:27 | pypa/warehouse | https://api.github.com/repos/pypa/warehouse | closed | "Confirm the username" should be "Confirm your username" | HTML UX/UI good first issue help needed raised in user testing | On the modal when a user wants to delete their account. Currently we show this message:

"Confirm the username" should be "Confirm your username"
---
**Good First Issue**: This issue is good for first time contributors. If you've already contributed to Warehouse, please work on [another issue without this label](https://github.com/pypa/warehouse/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aopen+-label%3A%22good+first+issue%22) instead. If there is not a corresponding pull request for this issue, it is up for grabs. For directions for getting set up, see our [Getting Started Guide](https://warehouse.pypa.io/development/getting-started/). If you are working on this issue and have questions, please feel free to ask them here, [`#pypa-dev` on Freenode](https://webchat.freenode.net/?channels=%23pypa-dev), or the [pypa-dev mailing list](https://groups.google.com/forum/#!forum/pypa-dev). | 1.0 | "Confirm the username" should be "Confirm your username" - On the modal when a user wants to delete their account. Currently we show this message:

"Confirm the username" should be "Confirm your username"
---
**Good First Issue**: This issue is good for first time contributors. If you've already contributed to Warehouse, please work on [another issue without this label](https://github.com/pypa/warehouse/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aopen+-label%3A%22good+first+issue%22) instead. If there is not a corresponding pull request for this issue, it is up for grabs. For directions for getting set up, see our [Getting Started Guide](https://warehouse.pypa.io/development/getting-started/). If you are working on this issue and have questions, please feel free to ask them here, [`#pypa-dev` on Freenode](https://webchat.freenode.net/?channels=%23pypa-dev), or the [pypa-dev mailing list](https://groups.google.com/forum/#!forum/pypa-dev). | non_priority | confirm the username should be confirm your username on the modal when a user wants to delete their account currently we show this message confirm the username should be confirm your username good first issue this issue is good for first time contributors if you ve already contributed to warehouse please work on instead if there is not a corresponding pull request for this issue it is up for grabs for directions for getting set up see our if you are working on this issue and have questions please feel free to ask them here or the | 0 |
9,064 | 3,020,687,177 | IssuesEvent | 2015-07-31 09:41:31 | oSoc15/educal | https://api.github.com/repos/oSoc15/educal | closed | Create event: not all fields should be required | 3 - Ready To Test bug | Only the title field should be required when creating an event
<img width="381" alt="pop" src="https://cloud.githubusercontent.com/assets/400335/8969012/085c9584-3641-11e5-9748-550814397d0d.png">
| 1.0 | Create event: not all fields should be required - Only the title field should be required when creating an event
<img width="381" alt="pop" src="https://cloud.githubusercontent.com/assets/400335/8969012/085c9584-3641-11e5-9748-550814397d0d.png">
| non_priority | create event not all fields should be required only the title field should be required when creating an event img width alt pop src | 0 |
386,160 | 26,670,008,191 | IssuesEvent | 2023-01-26 09:26:25 | hasnainmakada-99/Open-Source-With-Hasnain | https://api.github.com/repos/hasnainmakada-99/Open-Source-With-Hasnain | closed | Add new resources in open source | documentation enhancement help wanted good first issue EddieHub:good-first-issue | Add new resources in open sources for beginners to contribute to it and also get started with it | 1.0 | Add new resources in open source - Add new resources in open sources for beginners to contribute to it and also get started with it | non_priority | add new resources in open source add new resources in open sources for beginners to contribute to it and also get started with it | 0 |
294,798 | 25,406,355,072 | IssuesEvent | 2022-11-22 15:33:50 | mgratschew/COMP.SE.200-Assigntment | https://api.github.com/repos/mgratschew/COMP.SE.200-Assigntment | closed | Unit tests for capitalize.js | Test | Test cases:
1. Should return capitalized text for a lowercase string
2. Should return capitalized text for a capitalized string
3. Should return an empty string for an empty string
4. Should return strings of numbers unaffected | 1.0 | Unit tests for capitalize.js - Test cases:
1. Should return capitalized text for a lowercase string
2. Should return capitalized text for a capitalized string
3. Should return an empty string for an empty string
4. Should return strings of numbers unaffected | non_priority | unit tests for capitalize js test cases should return capitalized text for a lowercase string should return capitalized text for a capitalized string should return an empty string for an empty string should return strings of numbers unaffected | 0 |
58,645 | 14,444,711,823 | IssuesEvent | 2020-12-07 21:41:24 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | fatal error: too many errors emitted, stopping now [-ferror-limit=] 20 errors generated. | TF 2.2 subtype:macOS type:build/install | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): macOS Catalina version 10.15.4
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): from source
- TensorFlow version: 2.2
- Python version: 3.7
- Installed using virtualenv? pip? conda?: conda
- Bazel version (if compiling from source): 2.0.0
- GCC/Compiler version (if compiling from source): Apple clang version 11.0.3 (clang-1103.0.32.29)
- CUDA/cuDNN version: N/A
- GPU model and memory: N/A
**Describe the problem**
I'm trying to build tensorflow from source on macOS , within less than a few minutes, I get a sequence of c++ errors and the build fails
```
ERROR: /private/var/tmp/_bazel_emadboctor/e4710f39cfc38993e5fe8f0eb15a51f2/external/com_google_p
rotobuf/BUILD:148:1: C++ compilation of rule '@com_google_protobuf//:protobuf' failed (Exit 1)
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:68:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_BoolValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:82:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_BytesValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:96:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_DoubleValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:110:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_FloatValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:124:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_Int32Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:138:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_Int64Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:152:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_StringValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:166:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_UInt32Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:180:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_UInt64Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:310:20: error: no class named 'HasBitSetters' in 'google::protobuf::DoubleValue'
class DoubleValue::HasBitSetters {
~~~~~~~~~~~~~^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:315:24: error: redefinition of 'kValueFieldNumber' as different kind of symbol
const int DoubleValue::kValueFieldNumber;
^
/usr/local/include/google/protobuf/wrappers.pb.h:227:5: note: previous definition is here
kValueFieldNumber = 1,
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:412:19: error: out-of-line definition of 'MergePartialFromCodedStream' does not match any declaration in 'google::protobuf::DoubleValue'
bool DoubleValue::MergePartialFromCodedStream(
^~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:456:19: error: out-of-line definition of 'SerializeWithCachedSizes' does not match any declaration in 'google::protobuf::DoubleValue'
void DoubleValue::SerializeWithCachedSizes(
^~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:474:46: error: out-of-line definition of 'InternalSerializeWithCachedSizesToArray' does not match any declaration in 'google::protobuf::DoubleValue'
::PROTOBUF_NAMESPACE_ID::uint8* DoubleValue::InternalSerializeWithCachedSizesToArray(
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:561:19: error: redefinition of 'Swap'
void DoubleValue::Swap(DoubleValue* other) {
^
/usr/local/include/google/protobuf/wrappers.pb.h:154:15: note: previous definition is here
inline void Swap(DoubleValue* other) {
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:575:19: error: redefinition of 'UnsafeArenaSwap'
void DoubleValue::UnsafeArenaSwap(DoubleValue* other) {
^
/usr/local/include/google/protobuf/wrappers.pb.h:162:8: note: previous definition is here
void UnsafeArenaSwap(DoubleValue* other) {
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:595:19: error: no class named 'HasBitSetters' in 'google::protobuf::FloatValue'
class FloatValue::HasBitSetters {
~~~~~~~~~~~~^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:600:23: error: redefinition of 'kValueFieldNumber' as different kind of symbol
const int FloatValue::kValueFieldNumber;
^
/usr/local/include/google/protobuf/wrappers.pb.h:378:5: note: previous definition is here
kValueFieldNumber = 1,
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:697:18: error: out-of-line definition of 'MergePartialFromCodedStream' does not match any declaration in 'google::protobuf::FloatValue'
bool FloatValue::MergePartialFromCodedStream(
^~~~~~~~~~~~~~~~~~~~~~~~~~~
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
Target //tensorflow/tools/pip_package:build_pip_package failed to build
ERROR: /Users/emadboctor/Desktop/tensorflow/tensorflow/core/framework/BUILD:1084:1 C++ compilation of rule '@com_google_protobuf//:protobuf' failed (Exit 1)
INFO: Elapsed time: 58.678s, Critical Path: 7.64s
INFO: 213 processes: 213 local.
FAILED: Build did NOT complete successfully
(tfb) emadboctor@MacBook-Pro tensorflow %
```
**Provide the exact sequence of commands / steps that you executed before running into the problem**
```
conda create -n tfb
conda activate tfb
conda install pip six numpy wheel setuptools mock 'future>=0.17.1'
pip install -U keras_applications --no-deps
pip install -U keras_preprocessing --no-deps
git clone https://github.com/tensorflow/tensorflow.git
wget https://github.com/bazelbuild/bazel/releases/download/2.0.0/bazel-2.0.0-installer-darwin-x86_64.sh
chmod +x bazel-2.0.0-installer-darwin-x86_64.sh
./bazel-2.0.0-installer-darwin-x86_64.sh --user
export PATH=/home/emadboctor/bin:$PATH
cd tensorflow
./configure
You have bazel 2.0.0 installed.
Please specify the location of python. [Default is /Users/emadboctor/anaconda3/envs/tfb/bin/python]:
Found possible Python library paths:
/Users/emadboctor/anaconda3/envs/tfb/lib/python3.8/site-packages
Please input the desired Python library path to use. Default is [/Users/emadboctor/anaconda3/envs/tfb/lib/python3.8/site-packages]
Do you wish to build TensorFlow with OpenCL SYCL support? [y/N]:
No OpenCL SYCL support will be enabled for TensorFlow.
Do you wish to build TensorFlow with ROCm support? [y/N]:
No ROCm support will be enabled for TensorFlow.
Do you wish to build TensorFlow with CUDA support? [y/N]:
No CUDA support will be enabled for TensorFlow.
Do you wish to download a fresh release of clang? (Experimental) [y/N]: y
Clang will be downloaded and used to compile tensorflow.
Please specify optimization flags to use during compilation when bazel option "--config=opt" is specified [Default is -march=native -Wno-sign-compare]:
Would you like to interactively configure ./WORKSPACE for Android builds? [y/N]:
Not configuring the WORKSPACE for Android builds.
Do you wish to build TensorFlow with iOS support? [y/N]: y
iOS support will be enabled for TensorFlow.
Preconfigured Bazel build configs. You can use any of the below by adding "--config=<>" to your build command. See .bazelrc for more details.
--config=mkl # Build with MKL support.
--config=monolithic # Config for mostly static monolithic build.
--config=ngraph # Build with Intel nGraph support.
--config=numa # Build with NUMA support.
--config=dynamic_kernels # (Experimental) Build kernels into separate shared objects.
--config=v2 # Build TensorFlow 2.x instead of 1.x.
Preconfigured Bazel build configs to DISABLE default on features:
--config=noaws # Disable AWS S3 filesystem support.
--config=nogcp # Disable GCP support.
--config=nohdfs # Disable HDFS support.
--config=nonccl # Disable NVIDIA NCCL support.
Configuration finished
bazel build -c opt --config=mkl --copt=-mavx --copt=-mavx2 --copt=-mfma --copt=-msse4.2 --copt=-msse4.1 --jobs 8 --verbose_failures //tensorflow/tools/pip_package:build_pip_package
```
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| 1.0 | fatal error: too many errors emitted, stopping now [-ferror-limit=] 20 errors generated. - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): macOS Catalina version 10.15.4
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): from source
- TensorFlow version: 2.2
- Python version: 3.7
- Installed using virtualenv? pip? conda?: conda
- Bazel version (if compiling from source): 2.0.0
- GCC/Compiler version (if compiling from source): Apple clang version 11.0.3 (clang-1103.0.32.29)
- CUDA/cuDNN version: N/A
- GPU model and memory: N/A
**Describe the problem**
I'm trying to build tensorflow from source on macOS , within less than a few minutes, I get a sequence of c++ errors and the build fails
```
ERROR: /private/var/tmp/_bazel_emadboctor/e4710f39cfc38993e5fe8f0eb15a51f2/external/com_google_p
rotobuf/BUILD:148:1: C++ compilation of rule '@com_google_protobuf//:protobuf' failed (Exit 1)
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:68:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_BoolValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:82:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_BytesValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:96:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_DoubleValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:110:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_FloatValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:124:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_Int32Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:138:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_Int64Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:152:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_StringValue_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:166:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_UInt32Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:180:91: error: cannot initialize a member subobject of type 'int' with an lvalue of type 'void ()'
{{ATOMIC_VAR_INIT(::PROTOBUF_NAMESPACE_ID::internal::SCCInfoBase::kUninitialized), 0, InitDefaultsscc_info_UInt64Value_google_2fprotobuf_2fwrappers_2eproto}, {}};
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:310:20: error: no class named 'HasBitSetters' in 'google::protobuf::DoubleValue'
class DoubleValue::HasBitSetters {
~~~~~~~~~~~~~^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:315:24: error: redefinition of 'kValueFieldNumber' as different kind of symbol
const int DoubleValue::kValueFieldNumber;
^
/usr/local/include/google/protobuf/wrappers.pb.h:227:5: note: previous definition is here
kValueFieldNumber = 1,
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:412:19: error: out-of-line definition of 'MergePartialFromCodedStream' does not match any declaration in 'google::protobuf::DoubleValue'
bool DoubleValue::MergePartialFromCodedStream(
^~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:456:19: error: out-of-line definition of 'SerializeWithCachedSizes' does not match any declaration in 'google::protobuf::DoubleValue'
void DoubleValue::SerializeWithCachedSizes(
^~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:474:46: error: out-of-line definition of 'InternalSerializeWithCachedSizesToArray' does not match any declaration in 'google::protobuf::DoubleValue'
::PROTOBUF_NAMESPACE_ID::uint8* DoubleValue::InternalSerializeWithCachedSizesToArray(
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:561:19: error: redefinition of 'Swap'
void DoubleValue::Swap(DoubleValue* other) {
^
/usr/local/include/google/protobuf/wrappers.pb.h:154:15: note: previous definition is here
inline void Swap(DoubleValue* other) {
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:575:19: error: redefinition of 'UnsafeArenaSwap'
void DoubleValue::UnsafeArenaSwap(DoubleValue* other) {
^
/usr/local/include/google/protobuf/wrappers.pb.h:162:8: note: previous definition is here
void UnsafeArenaSwap(DoubleValue* other) {
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:595:19: error: no class named 'HasBitSetters' in 'google::protobuf::FloatValue'
class FloatValue::HasBitSetters {
~~~~~~~~~~~~^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:600:23: error: redefinition of 'kValueFieldNumber' as different kind of symbol
const int FloatValue::kValueFieldNumber;
^
/usr/local/include/google/protobuf/wrappers.pb.h:378:5: note: previous definition is here
kValueFieldNumber = 1,
^
external/com_google_protobuf/src/google/protobuf/wrappers.pb.cc:697:18: error: out-of-line definition of 'MergePartialFromCodedStream' does not match any declaration in 'google::protobuf::FloatValue'
bool FloatValue::MergePartialFromCodedStream(
^~~~~~~~~~~~~~~~~~~~~~~~~~~
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
Target //tensorflow/tools/pip_package:build_pip_package failed to build
ERROR: /Users/emadboctor/Desktop/tensorflow/tensorflow/core/framework/BUILD:1084:1 C++ compilation of rule '@com_google_protobuf//:protobuf' failed (Exit 1)
INFO: Elapsed time: 58.678s, Critical Path: 7.64s
INFO: 213 processes: 213 local.
FAILED: Build did NOT complete successfully
(tfb) emadboctor@MacBook-Pro tensorflow %
```
**Provide the exact sequence of commands / steps that you executed before running into the problem**
```
conda create -n tfb
conda activate tfb
conda install pip six numpy wheel setuptools mock 'future>=0.17.1'
pip install -U keras_applications --no-deps
pip install -U keras_preprocessing --no-deps
git clone https://github.com/tensorflow/tensorflow.git
wget https://github.com/bazelbuild/bazel/releases/download/2.0.0/bazel-2.0.0-installer-darwin-x86_64.sh
chmod +x bazel-2.0.0-installer-darwin-x86_64.sh
./bazel-2.0.0-installer-darwin-x86_64.sh --user
export PATH=/home/emadboctor/bin:$PATH
cd tensorflow
./configure
You have bazel 2.0.0 installed.
Please specify the location of python. [Default is /Users/emadboctor/anaconda3/envs/tfb/bin/python]:
Found possible Python library paths:
/Users/emadboctor/anaconda3/envs/tfb/lib/python3.8/site-packages
Please input the desired Python library path to use. Default is [/Users/emadboctor/anaconda3/envs/tfb/lib/python3.8/site-packages]
Do you wish to build TensorFlow with OpenCL SYCL support? [y/N]:
No OpenCL SYCL support will be enabled for TensorFlow.
Do you wish to build TensorFlow with ROCm support? [y/N]:
No ROCm support will be enabled for TensorFlow.
Do you wish to build TensorFlow with CUDA support? [y/N]:
No CUDA support will be enabled for TensorFlow.
Do you wish to download a fresh release of clang? (Experimental) [y/N]: y
Clang will be downloaded and used to compile tensorflow.
Please specify optimization flags to use during compilation when bazel option "--config=opt" is specified [Default is -march=native -Wno-sign-compare]:
Would you like to interactively configure ./WORKSPACE for Android builds? [y/N]:
Not configuring the WORKSPACE for Android builds.
Do you wish to build TensorFlow with iOS support? [y/N]: y
iOS support will be enabled for TensorFlow.
Preconfigured Bazel build configs. You can use any of the below by adding "--config=<>" to your build command. See .bazelrc for more details.
--config=mkl # Build with MKL support.
--config=monolithic # Config for mostly static monolithic build.
--config=ngraph # Build with Intel nGraph support.
--config=numa # Build with NUMA support.
--config=dynamic_kernels # (Experimental) Build kernels into separate shared objects.
--config=v2 # Build TensorFlow 2.x instead of 1.x.
Preconfigured Bazel build configs to DISABLE default on features:
--config=noaws # Disable AWS S3 filesystem support.
--config=nogcp # Disable GCP support.
--config=nohdfs # Disable HDFS support.
--config=nonccl # Disable NVIDIA NCCL support.
Configuration finished
bazel build -c opt --config=mkl --copt=-mavx --copt=-mavx2 --copt=-mfma --copt=-msse4.2 --copt=-msse4.1 --jobs 8 --verbose_failures //tensorflow/tools/pip_package:build_pip_package
```
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| non_priority | fatal error too many errors emitted stopping now errors generated please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu macos catalina version mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary from source tensorflow version python version installed using virtualenv pip conda conda bazel version if compiling from source gcc compiler version if compiling from source apple clang version clang cuda cudnn version n a gpu model and memory n a describe the problem i m trying to build tensorflow from source on macos within less than a few minutes i get a sequence of c errors and the build fails error private var tmp bazel emadboctor external com google p rotobuf build c compilation of rule com google protobuf protobuf failed exit external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info boolvalue google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info bytesvalue google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info doublevalue google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info floatvalue google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info stringvalue google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info google external com google protobuf src google protobuf wrappers pb cc error cannot initialize a member subobject of type int with an lvalue of type void atomic var init protobuf namespace id internal sccinfobase kuninitialized initdefaultsscc info google external com google protobuf src google protobuf wrappers pb cc error no class named hasbitsetters in google protobuf doublevalue class doublevalue hasbitsetters external com google protobuf src google protobuf wrappers pb cc error redefinition of kvaluefieldnumber as different kind of symbol const int doublevalue kvaluefieldnumber usr local include google protobuf wrappers pb h note previous definition is here kvaluefieldnumber external com google protobuf src google protobuf wrappers pb cc error out of line definition of mergepartialfromcodedstream does not match any declaration in google protobuf doublevalue bool doublevalue mergepartialfromcodedstream external com google protobuf src google protobuf wrappers pb cc error out of line definition of serializewithcachedsizes does not match any declaration in google protobuf doublevalue void doublevalue serializewithcachedsizes external com google protobuf src google protobuf wrappers pb cc error out of line definition of internalserializewithcachedsizestoarray does not match any declaration in google protobuf doublevalue protobuf namespace id doublevalue internalserializewithcachedsizestoarray external com google protobuf src google protobuf wrappers pb cc error redefinition of swap void doublevalue swap doublevalue other usr local include google protobuf wrappers pb h note previous definition is here inline void swap doublevalue other external com google protobuf src google protobuf wrappers pb cc error redefinition of unsafearenaswap void doublevalue unsafearenaswap doublevalue other usr local include google protobuf wrappers pb h note previous definition is here void unsafearenaswap doublevalue other external com google protobuf src google protobuf wrappers pb cc error no class named hasbitsetters in google protobuf floatvalue class floatvalue hasbitsetters external com google protobuf src google protobuf wrappers pb cc error redefinition of kvaluefieldnumber as different kind of symbol const int floatvalue kvaluefieldnumber usr local include google protobuf wrappers pb h note previous definition is here kvaluefieldnumber external com google protobuf src google protobuf wrappers pb cc error out of line definition of mergepartialfromcodedstream does not match any declaration in google protobuf floatvalue bool floatvalue mergepartialfromcodedstream fatal error too many errors emitted stopping now errors generated target tensorflow tools pip package build pip package failed to build error users emadboctor desktop tensorflow tensorflow core framework build c compilation of rule com google protobuf protobuf failed exit info elapsed time critical path info processes local failed build did not complete successfully tfb emadboctor macbook pro tensorflow provide the exact sequence of commands steps that you executed before running into the problem conda create n tfb conda activate tfb conda install pip six numpy wheel setuptools mock future pip install u keras applications no deps pip install u keras preprocessing no deps git clone wget chmod x bazel installer darwin sh bazel installer darwin sh user export path home emadboctor bin path cd tensorflow configure you have bazel installed please specify the location of python found possible python library paths users emadboctor envs tfb lib site packages please input the desired python library path to use default is do you wish to build tensorflow with opencl sycl support no opencl sycl support will be enabled for tensorflow do you wish to build tensorflow with rocm support no rocm support will be enabled for tensorflow do you wish to build tensorflow with cuda support no cuda support will be enabled for tensorflow do you wish to download a fresh release of clang experimental y clang will be downloaded and used to compile tensorflow please specify optimization flags to use during compilation when bazel option config opt is specified would you like to interactively configure workspace for android builds not configuring the workspace for android builds do you wish to build tensorflow with ios support y ios support will be enabled for tensorflow preconfigured bazel build configs you can use any of the below by adding config to your build command see bazelrc for more details config mkl build with mkl support config monolithic config for mostly static monolithic build config ngraph build with intel ngraph support config numa build with numa support config dynamic kernels experimental build kernels into separate shared objects config build tensorflow x instead of x preconfigured bazel build configs to disable default on features config noaws disable aws filesystem support config nogcp disable gcp support config nohdfs disable hdfs support config nonccl disable nvidia nccl support configuration finished bazel build c opt config mkl copt mavx copt copt mfma copt copt jobs verbose failures tensorflow tools pip package build pip package any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached | 0 |
298,665 | 22,549,224,164 | IssuesEvent | 2022-06-27 02:26:57 | Liqwid-Labs/liqwid-plutarch-extra | https://api.github.com/repos/Liqwid-Labs/liqwid-plutarch-extra | closed | Add construction/deconstruction of `PTagged` values to haddocks | documentation | We could use some usage guidelines/examples for (de)constructing ptagged values.
- Construction seems like it should be done with `pconstant`; does `pcon` work as well? Also seems like we typically need type annotations or application; I don't have a strong preference, but we should make a decision on what we want to consider best practice.
- Deconstruction can happen via `pmatch`, `pto`, or `pextract`. No preference here either, but we should pick one and add it.
We should also try to verify that the performance of each of these are the same. | 1.0 | Add construction/deconstruction of `PTagged` values to haddocks - We could use some usage guidelines/examples for (de)constructing ptagged values.
- Construction seems like it should be done with `pconstant`; does `pcon` work as well? Also seems like we typically need type annotations or application; I don't have a strong preference, but we should make a decision on what we want to consider best practice.
- Deconstruction can happen via `pmatch`, `pto`, or `pextract`. No preference here either, but we should pick one and add it.
We should also try to verify that the performance of each of these are the same. | non_priority | add construction deconstruction of ptagged values to haddocks we could use some usage guidelines examples for de constructing ptagged values construction seems like it should be done with pconstant does pcon work as well also seems like we typically need type annotations or application i don t have a strong preference but we should make a decision on what we want to consider best practice deconstruction can happen via pmatch pto or pextract no preference here either but we should pick one and add it we should also try to verify that the performance of each of these are the same | 0 |
214,305 | 24,062,863,532 | IssuesEvent | 2022-09-17 03:55:01 | SinLess-Games/Website | https://api.github.com/repos/SinLess-Games/Website | opened | h2-1.4.200.jar: 3 vulnerabilities (highest severity is: 9.8) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.200.jar</b></p></summary>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-42392](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42392) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | h2-1.4.200.jar | Direct | 2.0.206 | ❌ |
| [CVE-2022-23221](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23221) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | h2-1.4.200.jar | Direct | 2.1.210 | ❌ |
| [CVE-2021-23463](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23463) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | h2-1.4.200.jar | Direct | 2.0.202 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-42392</summary>
### Vulnerable Library - <b>h2-1.4.200.jar</b></p>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The org.h2.util.JdbcUtils.getConnection method of the H2 database takes as parameters the class name of the driver and URL of the database. An attacker may pass a JNDI driver name and a URL leading to a LDAP or RMI servers, causing remote code execution. This can be exploited through various attack vectors, most notably through the H2 Console which leads to unauthenticated remote code execution.
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42392>CVE-2021-42392</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/h2database/h2database/security/advisories/GHSA-h376-j262-vhq6">https://github.com/h2database/h2database/security/advisories/GHSA-h376-j262-vhq6</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution: 2.0.206</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-23221</summary>
### Vulnerable Library - <b>h2-1.4.200.jar</b></p>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
H2 Console before 2.1.210 allows remote attackers to execute arbitrary code via a jdbc:h2:mem JDBC URL containing the IGNORE_UNKNOWN_SETTINGS=TRUE;FORBID_CREATION=FALSE;INIT=RUNSCRIPT substring, a different vulnerability than CVE-2021-42392.
<p>Publish Date: 2022-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23221>CVE-2022-23221</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-01-19</p>
<p>Fix Resolution: 2.1.210</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23463</summary>
### Vulnerable Library - <b>h2-1.4.200.jar</b></p>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package com.h2database:h2 from 1.4.198 and before 2.0.202 are vulnerable to XML External Entity (XXE) Injection via the org.h2.jdbc.JdbcSQLXML class object, when it receives parsed string data from org.h2.jdbc.JdbcResultSet.getSQLXML() method. If it executes the getSource() method when the parameter is DOMSource.class it will trigger the vulnerability.
<p>Publish Date: 2021-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23463>CVE-2021-23463</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-12-10</p>
<p>Fix Resolution: 2.0.202</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | h2-1.4.200.jar: 3 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.200.jar</b></p></summary>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-42392](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42392) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | h2-1.4.200.jar | Direct | 2.0.206 | ❌ |
| [CVE-2022-23221](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23221) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | h2-1.4.200.jar | Direct | 2.1.210 | ❌ |
| [CVE-2021-23463](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23463) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | h2-1.4.200.jar | Direct | 2.0.202 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-42392</summary>
### Vulnerable Library - <b>h2-1.4.200.jar</b></p>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The org.h2.util.JdbcUtils.getConnection method of the H2 database takes as parameters the class name of the driver and URL of the database. An attacker may pass a JNDI driver name and a URL leading to a LDAP or RMI servers, causing remote code execution. This can be exploited through various attack vectors, most notably through the H2 Console which leads to unauthenticated remote code execution.
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42392>CVE-2021-42392</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/h2database/h2database/security/advisories/GHSA-h376-j262-vhq6">https://github.com/h2database/h2database/security/advisories/GHSA-h376-j262-vhq6</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution: 2.0.206</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-23221</summary>
### Vulnerable Library - <b>h2-1.4.200.jar</b></p>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
H2 Console before 2.1.210 allows remote attackers to execute arbitrary code via a jdbc:h2:mem JDBC URL containing the IGNORE_UNKNOWN_SETTINGS=TRUE;FORBID_CREATION=FALSE;INIT=RUNSCRIPT substring, a different vulnerability than CVE-2021-42392.
<p>Publish Date: 2022-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23221>CVE-2022-23221</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-01-19</p>
<p>Fix Resolution: 2.1.210</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23463</summary>
### Vulnerable Library - <b>h2-1.4.200.jar</b></p>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SinLess-Games/Website/commit/c5978132c1335273081c4b36a36c35bbbc597e2e">c5978132c1335273081c4b36a36c35bbbc597e2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package com.h2database:h2 from 1.4.198 and before 2.0.202 are vulnerable to XML External Entity (XXE) Injection via the org.h2.jdbc.JdbcSQLXML class object, when it receives parsed string data from org.h2.jdbc.JdbcResultSet.getSQLXML() method. If it executes the getSource() method when the parameter is DOMSource.class it will trigger the vulnerability.
<p>Publish Date: 2021-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23463>CVE-2021-23463</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-12-10</p>
<p>Fix Resolution: 2.0.202</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_priority | jar vulnerabilities highest severity is vulnerable library jar database engine library home page a href path to dependency file build gradle path to vulnerable library gradle caches modules files com jar home wss scanner gradle caches modules files com jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high jar direct high jar direct high jar direct details cve vulnerable library jar database engine library home page a href path to dependency file build gradle path to vulnerable library gradle caches modules files com jar home wss scanner gradle caches modules files com jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details the org util jdbcutils getconnection method of the database takes as parameters the class name of the driver and url of the database an attacker may pass a jndi driver name and a url leading to a ldap or rmi servers causing remote code execution this can be exploited through various attack vectors most notably through the console which leads to unauthenticated remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend cve vulnerable library jar database engine library home page a href path to dependency file build gradle path to vulnerable library gradle caches modules files com jar home wss scanner gradle caches modules files com jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details console before allows remote attackers to execute arbitrary code via a jdbc mem jdbc url containing the ignore unknown settings true forbid creation false init runscript substring a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend cve vulnerable library jar database engine library home page a href path to dependency file build gradle path to vulnerable library gradle caches modules files com jar home wss scanner gradle caches modules files com jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details the package com from and before are vulnerable to xml external entity xxe injection via the org jdbc jdbcsqlxml class object when it receives parsed string data from org jdbc jdbcresultset getsqlxml method if it executes the getsource method when the parameter is domsource class it will trigger the vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend | 0 |
261,374 | 19,708,657,934 | IssuesEvent | 2022-01-13 01:49:04 | bluesky/tiled | https://api.github.com/repos/bluesky/tiled | opened | Add guide for serving related files | documentation | I think that https://blueskyproject.io/tiled/explanations/scaling-down.html may be hard to follow (I wrote it, so I’m critiquing my own work here) and it would be better to replace it with a how-to guide with a practical example involving files with the same basename that should be represented by one node with children. | 1.0 | Add guide for serving related files - I think that https://blueskyproject.io/tiled/explanations/scaling-down.html may be hard to follow (I wrote it, so I’m critiquing my own work here) and it would be better to replace it with a how-to guide with a practical example involving files with the same basename that should be represented by one node with children. | non_priority | add guide for serving related files i think that may be hard to follow i wrote it so i’m critiquing my own work here and it would be better to replace it with a how to guide with a practical example involving files with the same basename that should be represented by one node with children | 0 |
53,388 | 13,161,423,622 | IssuesEvent | 2020-08-10 19:30:40 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | libtensorflowlite.so built with flex delegates too big! | TFLiteConverter stat:awaiting response type:build/install | **System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04 x86_64 GNU/Linux Docker (official tensorflow image)
- TensorFlow installed from (source or binary): pip
- TensorFlow version (or github SHA if from source): 2.4.0-dev20200712
Modified BUILD to add following dependency:
```
"//tensorflow/lite/delegates/flex:delegate",
```
And did a build using the following command:
```
bazel build --config=monolithic --define=with_select_tf_ops=true -c opt //tensorflow/lite:libtensorflowlite.so
```
**The output:**
`libtensorflowlite.so` that is `147104080` bytes (147MB)
Followed the above instructions to generate a libtensorflow with interpreter support for flex ops. Is this file size of 147MB on the library expected? Is this because I added the flex:delegate dependency? How can I reduce this to be <10MB like the lib produced if I follow the lite/tools/make shell scripts (i.e build_lib.sh) | 1.0 | libtensorflowlite.so built with flex delegates too big! - **System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04 x86_64 GNU/Linux Docker (official tensorflow image)
- TensorFlow installed from (source or binary): pip
- TensorFlow version (or github SHA if from source): 2.4.0-dev20200712
Modified BUILD to add following dependency:
```
"//tensorflow/lite/delegates/flex:delegate",
```
And did a build using the following command:
```
bazel build --config=monolithic --define=with_select_tf_ops=true -c opt //tensorflow/lite:libtensorflowlite.so
```
**The output:**
`libtensorflowlite.so` that is `147104080` bytes (147MB)
Followed the above instructions to generate a libtensorflow with interpreter support for flex ops. Is this file size of 147MB on the library expected? Is this because I added the flex:delegate dependency? How can I reduce this to be <10MB like the lib produced if I follow the lite/tools/make shell scripts (i.e build_lib.sh) | non_priority | libtensorflowlite so built with flex delegates too big system information os platform and distribution e g linux ubuntu ubuntu gnu linux docker official tensorflow image tensorflow installed from source or binary pip tensorflow version or github sha if from source modified build to add following dependency tensorflow lite delegates flex delegate and did a build using the following command bazel build config monolithic define with select tf ops true c opt tensorflow lite libtensorflowlite so the output libtensorflowlite so that is bytes followed the above instructions to generate a libtensorflow with interpreter support for flex ops is this file size of on the library expected is this because i added the flex delegate dependency how can i reduce this to be like the lib produced if i follow the lite tools make shell scripts i e build lib sh | 0 |
217,274 | 16,848,848,478 | IssuesEvent | 2021-06-20 04:17:17 | hakehuang/infoflow | https://api.github.com/repos/hakehuang/infoflow | opened |
tests-ci :kernel.memory_protection.create_new_thread_from_user_invalid_stacksize : zephyr-v2.6.0-286-g46029914a7ac: lpcxpresso55s28: test Timeout
| area: Tests |
**Describe the bug**
kernel.memory_protection.create_new_thread_from_user_invalid_stacksize test is Timeout on zephyr-v2.6.0-286-g46029914a7ac on lpcxpresso55s28
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p lpcxpresso55s28 --testcase-root tests --sub-test kernel.memory_protection
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
-
*** Booting Zephyr OS build zephyr-v2.6.0-286-g46029914a7ac ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| 1.0 |
tests-ci :kernel.memory_protection.create_new_thread_from_user_invalid_stacksize : zephyr-v2.6.0-286-g46029914a7ac: lpcxpresso55s28: test Timeout
-
**Describe the bug**
kernel.memory_protection.create_new_thread_from_user_invalid_stacksize test is Timeout on zephyr-v2.6.0-286-g46029914a7ac on lpcxpresso55s28
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p lpcxpresso55s28 --testcase-root tests --sub-test kernel.memory_protection
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
-
*** Booting Zephyr OS build zephyr-v2.6.0-286-g46029914a7ac ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| non_priority | tests ci kernel memory protection create new thread from user invalid stacksize zephyr test timeout describe the bug kernel memory protection create new thread from user invalid stacksize test is timeout on zephyr on see logs for details to reproduce scripts twister device testing device serial dev p testcase root tests sub test kernel memory protection see error expected behavior test pass impact logs and console output booting zephyr os build zephyr running test suite memory protection test suite start test permission inheritance assertion fail west topdir zephyr arch arm core cortex m fault c esf could not be retrieved successfully shall never occur assertion fail west topdir zephyr arch arm core cortex m fault c esf could not be retrieved successfully shall never occur environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr | 0 |
111,024 | 17,009,691,976 | IssuesEvent | 2021-07-02 01:07:40 | tamirdahan/keycloak | https://api.github.com/repos/tamirdahan/keycloak | opened | CVE-2012-0881 (High) detected in xercesImpl-2.9.1-jbossas-2.jar, xercesImpl-2.11.0.jar | security vulnerability | ## CVE-2012-0881 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>xercesImpl-2.9.1-jbossas-2.jar</b>, <b>xercesImpl-2.11.0.jar</b></p></summary>
<p>
<details><summary><b>xercesImpl-2.9.1-jbossas-2.jar</b></p></summary>
<p>Apache Xerces</p>
<p>Library home page: <a href="https://xerces.apache.org/xerces2-j/">https://xerces.apache.org/xerces2-j/</a></p>
<p>Path to dependency file: keycloak/adapters/saml/as7-eap6/subsystem/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1-jbossas-2/xercesImpl-2.9.1-jbossas-2.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1-jbossas-2/xercesImpl-2.9.1-jbossas-2.jar</p>
<p>
Dependency Hierarchy:
- keycloak-saml-as7-adapter-15.0.0-SNAPSHOT.jar (Root Library)
- keycloak-saml-tomcat-adapter-core-15.0.0-SNAPSHOT.jar
- keycloak-saml-core-15.0.0-SNAPSHOT.jar
- rt-2.3.1.jar
- saaj-impl-1.3.16-jbossorg-1.jar
- :x: **xercesImpl-2.9.1-jbossas-2.jar** (Vulnerable Library)
</details>
<details><summary><b>xercesImpl-2.11.0.jar</b></p></summary>
<p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI), a complete framework for building parser components and configurations that is extremely modular and easy to program.
The Apache Xerces2 parser is the reference implementation of XNI but other parser components, configurations, and parsers can be written using the Xerces Native Interface. For complete design and implementation documents, refer to the XNI Manual.
Xerces2 is a fully conforming XML Schema 1.0 processor. A partial experimental implementation of the XML Schema 1.1 Structures and Datatypes Working Drafts (December 2009) and an experimental implementation of the XML Schema Definition Language (XSD): Component Designators (SCD) Candidate Recommendation (January 2010) are provided for evaluation. For more information, refer to the XML Schema page.
Xerces2 also provides a complete implementation of the Document Object Model Level 3 Core and Load/Save W3C Recommendations and provides a complete implementation of the XML Inclusions (XInclude) W3C Recommendation. It also provides support for OASIS XML Catalogs v1.1.
Xerces2 is able to parse documents written according to the XML 1.1 Recommendation, except that it does not yet provide an option to enable normalization checking as described in section 2.13 of this specification. It also handles namespaces according to the XML Namespaces 1.1 Recommendation, and will correctly serialize XML 1.1 documents if the DOM level 3 load/save APIs are in use.</p>
<p>Library home page: <a href="https://xerces.apache.org/xerces2-j/">https://xerces.apache.org/xerces2-j/</a></p>
<p>Path to dependency file: keycloak/testsuite/model/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.11.0/xercesImpl-2.11.0.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.11.0/xercesImpl-2.11.0.jar</p>
<p>
Dependency Hierarchy:
- graphene-webdriver-2.3.2.pom (Root Library)
- arquillian-drone-webdriver-depchain-2.5.2.pom
- htmlunit-driver-2.27.jar
- htmlunit-2.27.jar
- neko-htmlunit-2.27.jar
- :x: **xercesImpl-2.11.0.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Xerces2 Java Parser before 2.12.0 allows remote attackers to cause a denial of service (CPU consumption) via a crafted message to an XML service, which triggers hash table collisions.
<p>Publish Date: 2017-10-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-0881>CVE-2012-0881</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0881">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0881</a></p>
<p>Release Date: 2017-10-30</p>
<p>Fix Resolution: 2.12.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"xerces","packageName":"xercesImpl","packageVersion":"2.9.1-jbossas-2","packageFilePaths":["/adapters/saml/as7-eap6/subsystem/pom.xml","/adapters/saml/as7-eap6/adapter/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.keycloak:keycloak-saml-as7-adapter:15.0.0-SNAPSHOT;org.keycloak:keycloak-saml-tomcat-adapter-core:15.0.0-SNAPSHOT;org.keycloak:keycloak-saml-core:15.0.0-SNAPSHOT;com.sun.xml.ws:rt:2.3.1;com.sun.xml.messaging.saaj:saaj-impl:1.3.16-jbossorg-1;xerces:xercesImpl:2.9.1-jbossas-2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.12.0"},{"packageType":"Java","groupId":"xerces","packageName":"xercesImpl","packageVersion":"2.11.0","packageFilePaths":["/testsuite/model/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.jboss.arquillian.graphene:graphene-webdriver:2.3.2;org.jboss.arquillian.extension:arquillian-drone-webdriver-depchain:2.5.2;org.seleniumhq.selenium:htmlunit-driver:2.27;net.sourceforge.htmlunit:htmlunit:2.27;net.sourceforge.htmlunit:neko-htmlunit:2.27;xerces:xercesImpl:2.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.12.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2012-0881","vulnerabilityDetails":"Apache Xerces2 Java Parser before 2.12.0 allows remote attackers to cause a denial of service (CPU consumption) via a crafted message to an XML service, which triggers hash table collisions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-0881","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2012-0881 (High) detected in xercesImpl-2.9.1-jbossas-2.jar, xercesImpl-2.11.0.jar - ## CVE-2012-0881 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>xercesImpl-2.9.1-jbossas-2.jar</b>, <b>xercesImpl-2.11.0.jar</b></p></summary>
<p>
<details><summary><b>xercesImpl-2.9.1-jbossas-2.jar</b></p></summary>
<p>Apache Xerces</p>
<p>Library home page: <a href="https://xerces.apache.org/xerces2-j/">https://xerces.apache.org/xerces2-j/</a></p>
<p>Path to dependency file: keycloak/adapters/saml/as7-eap6/subsystem/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1-jbossas-2/xercesImpl-2.9.1-jbossas-2.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1-jbossas-2/xercesImpl-2.9.1-jbossas-2.jar</p>
<p>
Dependency Hierarchy:
- keycloak-saml-as7-adapter-15.0.0-SNAPSHOT.jar (Root Library)
- keycloak-saml-tomcat-adapter-core-15.0.0-SNAPSHOT.jar
- keycloak-saml-core-15.0.0-SNAPSHOT.jar
- rt-2.3.1.jar
- saaj-impl-1.3.16-jbossorg-1.jar
- :x: **xercesImpl-2.9.1-jbossas-2.jar** (Vulnerable Library)
</details>
<details><summary><b>xercesImpl-2.11.0.jar</b></p></summary>
<p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI), a complete framework for building parser components and configurations that is extremely modular and easy to program.
The Apache Xerces2 parser is the reference implementation of XNI but other parser components, configurations, and parsers can be written using the Xerces Native Interface. For complete design and implementation documents, refer to the XNI Manual.
Xerces2 is a fully conforming XML Schema 1.0 processor. A partial experimental implementation of the XML Schema 1.1 Structures and Datatypes Working Drafts (December 2009) and an experimental implementation of the XML Schema Definition Language (XSD): Component Designators (SCD) Candidate Recommendation (January 2010) are provided for evaluation. For more information, refer to the XML Schema page.
Xerces2 also provides a complete implementation of the Document Object Model Level 3 Core and Load/Save W3C Recommendations and provides a complete implementation of the XML Inclusions (XInclude) W3C Recommendation. It also provides support for OASIS XML Catalogs v1.1.
Xerces2 is able to parse documents written according to the XML 1.1 Recommendation, except that it does not yet provide an option to enable normalization checking as described in section 2.13 of this specification. It also handles namespaces according to the XML Namespaces 1.1 Recommendation, and will correctly serialize XML 1.1 documents if the DOM level 3 load/save APIs are in use.</p>
<p>Library home page: <a href="https://xerces.apache.org/xerces2-j/">https://xerces.apache.org/xerces2-j/</a></p>
<p>Path to dependency file: keycloak/testsuite/model/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.11.0/xercesImpl-2.11.0.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.11.0/xercesImpl-2.11.0.jar</p>
<p>
Dependency Hierarchy:
- graphene-webdriver-2.3.2.pom (Root Library)
- arquillian-drone-webdriver-depchain-2.5.2.pom
- htmlunit-driver-2.27.jar
- htmlunit-2.27.jar
- neko-htmlunit-2.27.jar
- :x: **xercesImpl-2.11.0.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Xerces2 Java Parser before 2.12.0 allows remote attackers to cause a denial of service (CPU consumption) via a crafted message to an XML service, which triggers hash table collisions.
<p>Publish Date: 2017-10-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-0881>CVE-2012-0881</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0881">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0881</a></p>
<p>Release Date: 2017-10-30</p>
<p>Fix Resolution: 2.12.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"xerces","packageName":"xercesImpl","packageVersion":"2.9.1-jbossas-2","packageFilePaths":["/adapters/saml/as7-eap6/subsystem/pom.xml","/adapters/saml/as7-eap6/adapter/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.keycloak:keycloak-saml-as7-adapter:15.0.0-SNAPSHOT;org.keycloak:keycloak-saml-tomcat-adapter-core:15.0.0-SNAPSHOT;org.keycloak:keycloak-saml-core:15.0.0-SNAPSHOT;com.sun.xml.ws:rt:2.3.1;com.sun.xml.messaging.saaj:saaj-impl:1.3.16-jbossorg-1;xerces:xercesImpl:2.9.1-jbossas-2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.12.0"},{"packageType":"Java","groupId":"xerces","packageName":"xercesImpl","packageVersion":"2.11.0","packageFilePaths":["/testsuite/model/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.jboss.arquillian.graphene:graphene-webdriver:2.3.2;org.jboss.arquillian.extension:arquillian-drone-webdriver-depchain:2.5.2;org.seleniumhq.selenium:htmlunit-driver:2.27;net.sourceforge.htmlunit:htmlunit:2.27;net.sourceforge.htmlunit:neko-htmlunit:2.27;xerces:xercesImpl:2.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.12.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2012-0881","vulnerabilityDetails":"Apache Xerces2 Java Parser before 2.12.0 allows remote attackers to cause a denial of service (CPU consumption) via a crafted message to an XML service, which triggers hash table collisions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-0881","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in xercesimpl jbossas jar xercesimpl jar cve high severity vulnerability vulnerable libraries xercesimpl jbossas jar xercesimpl jar xercesimpl jbossas jar apache xerces library home page a href path to dependency file keycloak adapters saml subsystem pom xml path to vulnerable library home wss scanner repository xerces xercesimpl jbossas xercesimpl jbossas jar home wss scanner repository xerces xercesimpl jbossas xercesimpl jbossas jar dependency hierarchy keycloak saml adapter snapshot jar root library keycloak saml tomcat adapter core snapshot jar keycloak saml core snapshot jar rt jar saaj impl jbossorg jar x xercesimpl jbossas jar vulnerable library xercesimpl jar is the next generation of high performance fully compliant xml parsers in the apache xerces family this new version of xerces introduces the xerces native interface xni a complete framework for building parser components and configurations that is extremely modular and easy to program the apache parser is the reference implementation of xni but other parser components configurations and parsers can be written using the xerces native interface for complete design and implementation documents refer to the xni manual is a fully conforming xml schema processor a partial experimental implementation of the xml schema structures and datatypes working drafts december and an experimental implementation of the xml schema definition language xsd component designators scd candidate recommendation january are provided for evaluation for more information refer to the xml schema page also provides a complete implementation of the document object model level core and load save recommendations and provides a complete implementation of the xml inclusions xinclude recommendation it also provides support for oasis xml catalogs is able to parse documents written according to the xml recommendation except that it does not yet provide an option to enable normalization checking as described in section of this specification it also handles namespaces according to the xml namespaces recommendation and will correctly serialize xml documents if the dom level load save apis are in use library home page a href path to dependency file keycloak testsuite model pom xml path to vulnerable library home wss scanner repository xerces xercesimpl xercesimpl jar home wss scanner repository xerces xercesimpl xercesimpl jar dependency hierarchy graphene webdriver pom root library arquillian drone webdriver depchain pom htmlunit driver jar htmlunit jar neko htmlunit jar x xercesimpl jar vulnerable library found in base branch master vulnerability details apache java parser before allows remote attackers to cause a denial of service cpu consumption via a crafted message to an xml service which triggers hash table collisions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org keycloak keycloak saml adapter snapshot org keycloak keycloak saml tomcat adapter core snapshot org keycloak keycloak saml core snapshot com sun xml ws rt com sun xml messaging saaj saaj impl jbossorg xerces xercesimpl jbossas isminimumfixversionavailable true minimumfixversion packagetype java groupid xerces packagename xercesimpl packageversion packagefilepaths istransitivedependency true dependencytree org jboss arquillian graphene graphene webdriver org jboss arquillian extension arquillian drone webdriver depchain org seleniumhq selenium htmlunit driver net sourceforge htmlunit htmlunit net sourceforge htmlunit neko htmlunit xerces xercesimpl isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails apache java parser before allows remote attackers to cause a denial of service cpu consumption via a crafted message to an xml service which triggers hash table collisions vulnerabilityurl | 0 |
13,478 | 5,378,717,088 | IssuesEvent | 2017-02-23 15:36:07 | dotnet/docs | https://api.github.com/repos/dotnet/docs | opened | Unit Test tutorials need better description to get started | Goal - Update from project.json based projects to MSBuild based projects | Both of the unit testing tutorials for .NET Core start 'hot', without providing a good introduction on how to start working on the tutorial.
The affected files are https://github.com/dotnet/docs/blob/master/docs/core/testing/unit-testing-with-dotnet-test.md and https://github.com/dotnet/docs/blob/master/docs/core/testing/using-mstest-on-windows.md
This should be fixed on the csproj branch, as these are updated for the new project system.
The new introduction should show:
1. How to start by creating the working directory and getting ready to create the new projects as currently described.
2. How to explore the finished solution by downloading the code from the github repository.
It should also explain why you would choose one or the other: The first provides an interactive experience where readers can bulid the solution and see how the solution grows. The second enables a quicker path to see a full solutions.
Note: this is based on a LiveFyre comment on the dotnet test article.
| 1.0 | Unit Test tutorials need better description to get started - Both of the unit testing tutorials for .NET Core start 'hot', without providing a good introduction on how to start working on the tutorial.
The affected files are https://github.com/dotnet/docs/blob/master/docs/core/testing/unit-testing-with-dotnet-test.md and https://github.com/dotnet/docs/blob/master/docs/core/testing/using-mstest-on-windows.md
This should be fixed on the csproj branch, as these are updated for the new project system.
The new introduction should show:
1. How to start by creating the working directory and getting ready to create the new projects as currently described.
2. How to explore the finished solution by downloading the code from the github repository.
It should also explain why you would choose one or the other: The first provides an interactive experience where readers can bulid the solution and see how the solution grows. The second enables a quicker path to see a full solutions.
Note: this is based on a LiveFyre comment on the dotnet test article.
| non_priority | unit test tutorials need better description to get started both of the unit testing tutorials for net core start hot without providing a good introduction on how to start working on the tutorial the affected files are and this should be fixed on the csproj branch as these are updated for the new project system the new introduction should show how to start by creating the working directory and getting ready to create the new projects as currently described how to explore the finished solution by downloading the code from the github repository it should also explain why you would choose one or the other the first provides an interactive experience where readers can bulid the solution and see how the solution grows the second enables a quicker path to see a full solutions note this is based on a livefyre comment on the dotnet test article | 0 |
40,095 | 12,746,540,106 | IssuesEvent | 2020-06-26 16:07:25 | tech256/jobs | https://api.github.com/repos/tech256/jobs | closed | Senior Cyber Security Engineer | Cyber Security Hiring stale |
Senior Cyber Security Engineer
Vicksburg, MS
Are you a pro at Cyber Security? Do you enjoy supporting our military? INSUVI, Inc. is looking for great talent to join our team!
What We Can Offer YOU!
Medical
Dental
Vision
Long and Short-Term Disability
Life Insurance
401(k)
Paid Time Off (PTO)
Paid Holidays
And More!
COMPANY OVERVIEW:
INSUVI, Inc., is a certified Economically Disadvantaged Woman-Owned Small Business (EDWOSB) headquartered in Huntsville, Alabama. We provide Information Technology, JavaScript Training, Systems Engineering, and Training services.
POSITION OVERVIEW:
Job Responsibilities
Leads security event monitoring and security configuration of Palo Alto Firewall/IDPS, Cisco ASA, Sourcefire, FireEye, BRO, SNORT and similar intrusion detection and prevention technologies
Conducts research and documents threats and their behavior
Monitors internal and external cyber threat intelligence sources
Provides recommendations on threat mitigation strategies
Performs routine event reporting over time including trend reporting and analysis
Configures and maintains various cyber security platforms
Defines/maintains security configurations and policies for Intrusion Detection Systems (IDS) and Intrusion Prevention Systems (IPS) technologies
Maintains detection signatures; deploys new detection signatures
Monitors Security Information and Event Management (SIEM) events related to implemented IDS/IPS technologies
Configures and enforces audit and logging policies for IDS/IPS technologies
Defines/monitors Security Technical Implementation Guide (STIG) compliance of intrusion management technologies
QUALIFICATIONS:
Education & Experience
Bachelor's degree from an accredited university/college in Computer Science, Computer Engineering or related field and 8-14 years of prior relevant experience or master's degree with 6 -12 years of prior relevant experience
Relevant Experience required:
Experience required in security or network technology (Unix/Windows OS, Cisco/Juniper Routing-Switching) within a hands-on design/Implementation/Administration role
Knowledge & Skills
Ability to configure and support SIEM platforms like ArcSight, ELK or similar
Proven experience and ability to leverage CND analyst toolsets to detect and respond to IT security incidents
Ability to implement standard procedures for intrusion and related cyber incident response
Demonstrates in-depth knowledge of TCP-IP protocol implementations for all common network services
Professionally certified, within a CND discipline, as Technical Level III as defined by DODI 8570 is a requirement
Desired skills: Palo Alto Firewall and IDS/IPS, Sourcefire, FireEye, Arcsight/SIEM, JRSS, ASA, Linux System Administration; Win
Other Requirements
Certifications Required: CISSP, and PCNSE or CCNA-Security
Desired Certifications: CISSP-ISSAP, OSCP
INSUVI, Inc., provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, gender identity, sexual orientation, pregnancy, status as a parent, national origin, status as a parent, disability (physical or mental), family medical history or genetic information, political affiliation, military service, or other non-merit based factors.
For more information, or to apply now, you must go to the website below. Please DO NOT email your resume to us as we only accept applications through our website.
https://www.applicantpro.com/j/1311414-219641 | True | Senior Cyber Security Engineer -
Senior Cyber Security Engineer
Vicksburg, MS
Are you a pro at Cyber Security? Do you enjoy supporting our military? INSUVI, Inc. is looking for great talent to join our team!
What We Can Offer YOU!
Medical
Dental
Vision
Long and Short-Term Disability
Life Insurance
401(k)
Paid Time Off (PTO)
Paid Holidays
And More!
COMPANY OVERVIEW:
INSUVI, Inc., is a certified Economically Disadvantaged Woman-Owned Small Business (EDWOSB) headquartered in Huntsville, Alabama. We provide Information Technology, JavaScript Training, Systems Engineering, and Training services.
POSITION OVERVIEW:
Job Responsibilities
Leads security event monitoring and security configuration of Palo Alto Firewall/IDPS, Cisco ASA, Sourcefire, FireEye, BRO, SNORT and similar intrusion detection and prevention technologies
Conducts research and documents threats and their behavior
Monitors internal and external cyber threat intelligence sources
Provides recommendations on threat mitigation strategies
Performs routine event reporting over time including trend reporting and analysis
Configures and maintains various cyber security platforms
Defines/maintains security configurations and policies for Intrusion Detection Systems (IDS) and Intrusion Prevention Systems (IPS) technologies
Maintains detection signatures; deploys new detection signatures
Monitors Security Information and Event Management (SIEM) events related to implemented IDS/IPS technologies
Configures and enforces audit and logging policies for IDS/IPS technologies
Defines/monitors Security Technical Implementation Guide (STIG) compliance of intrusion management technologies
QUALIFICATIONS:
Education & Experience
Bachelor's degree from an accredited university/college in Computer Science, Computer Engineering or related field and 8-14 years of prior relevant experience or master's degree with 6 -12 years of prior relevant experience
Relevant Experience required:
Experience required in security or network technology (Unix/Windows OS, Cisco/Juniper Routing-Switching) within a hands-on design/Implementation/Administration role
Knowledge & Skills
Ability to configure and support SIEM platforms like ArcSight, ELK or similar
Proven experience and ability to leverage CND analyst toolsets to detect and respond to IT security incidents
Ability to implement standard procedures for intrusion and related cyber incident response
Demonstrates in-depth knowledge of TCP-IP protocol implementations for all common network services
Professionally certified, within a CND discipline, as Technical Level III as defined by DODI 8570 is a requirement
Desired skills: Palo Alto Firewall and IDS/IPS, Sourcefire, FireEye, Arcsight/SIEM, JRSS, ASA, Linux System Administration; Win
Other Requirements
Certifications Required: CISSP, and PCNSE or CCNA-Security
Desired Certifications: CISSP-ISSAP, OSCP
INSUVI, Inc., provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, gender identity, sexual orientation, pregnancy, status as a parent, national origin, status as a parent, disability (physical or mental), family medical history or genetic information, political affiliation, military service, or other non-merit based factors.
For more information, or to apply now, you must go to the website below. Please DO NOT email your resume to us as we only accept applications through our website.
https://www.applicantpro.com/j/1311414-219641 | non_priority | senior cyber security engineer senior cyber security engineer vicksburg ms are you a pro at cyber security do you enjoy supporting our military insuvi inc is looking for great talent to join our team what we can offer you medical dental vision long and short term disability life insurance k paid time off pto paid holidays and more company overview insuvi inc is a certified economically disadvantaged woman owned small business edwosb headquartered in huntsville alabama we provide information technology javascript training systems engineering and training services position overview job responsibilities leads security event monitoring and security configuration of palo alto firewall idps cisco asa sourcefire fireeye bro snort and similar intrusion detection and prevention technologies conducts research and documents threats and their behavior monitors internal and external cyber threat intelligence sources provides recommendations on threat mitigation strategies performs routine event reporting over time including trend reporting and analysis configures and maintains various cyber security platforms defines maintains security configurations and policies for intrusion detection systems ids and intrusion prevention systems ips technologies maintains detection signatures deploys new detection signatures monitors security information and event management siem events related to implemented ids ips technologies configures and enforces audit and logging policies for ids ips technologies defines monitors security technical implementation guide stig compliance of intrusion management technologies qualifications education experience bachelor s degree from an accredited university college in computer science computer engineering or related field and years of prior relevant experience or master s degree with years of prior relevant experience relevant experience required experience required in security or network technology unix windows os cisco juniper routing switching within a hands on design implementation administration role knowledge skills ability to configure and support siem platforms like arcsight elk or similar proven experience and ability to leverage cnd analyst toolsets to detect and respond to it security incidents ability to implement standard procedures for intrusion and related cyber incident response demonstrates in depth knowledge of tcp ip protocol implementations for all common network services professionally certified within a cnd discipline as technical level iii as defined by dodi is a requirement desired skills palo alto firewall and ids ips sourcefire fireeye arcsight siem jrss asa linux system administration win other requirements certifications required cissp and pcnse or ccna security desired certifications cissp issap oscp insuvi inc provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race color religion age sex gender identity sexual orientation pregnancy status as a parent national origin status as a parent disability physical or mental family medical history or genetic information political affiliation military service or other non merit based factors for more information or to apply now you must go to the website below please do not email your resume to us as we only accept applications through our website | 0 |
127,600 | 10,475,588,860 | IssuesEvent | 2019-09-23 16:39:39 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | opened | Manual test run on Linux for 0.69.x - Release | OS/unix-like/linux QA/Yes release-notes/exclude tests | ## Per release specialty tests
- [ ] Chromecast support.([#209](https://github.com/brave/brave-browser/issues/209))
- [ ] Add Dapp detection.([#718](https://github.com/brave/brave-browser/issues/718))
- [ ] Scale cover image height on tipping banner.([#2015](https://github.com/brave/brave-browser/issues/2015))
- [ ] Migrate sync code to Syncable Service.([#2754](https://github.com/brave/brave-browser/issues/2754))
- [ ] Cosmetic filter !important.([#3041](https://github.com/brave/brave-browser/issues/3041))
- [ ] Warning message not being displayed, when click on `Confirm Sync code` without any code added.([#3190](https://github.com/brave/brave-browser/issues/3190))
- [ ] Welcome screen: support dark mode .([#3911](https://github.com/brave/brave-browser/issues/3911))
- [ ] Implement user feedback loop.([#4047](https://github.com/brave/brave-browser/issues/4047))
- [ ] Blocked ads sometimes leave additional white space on page.([#4338](https://github.com/brave/brave-browser/issues/4338))
- [ ] Custom adblock rules should allow cosmetic filters .([#4348](https://github.com/brave/brave-browser/issues/4348))
- [ ] Add built-in Ethereum remote client / wallet / Dapp browser.([#4494](https://github.com/brave/brave-browser/issues/4494))
- [ ] Add a menu item for triggering WebUI for the Ethereum wallet.([#4579](https://github.com/brave/brave-browser/issues/4579))
- [ ] Settings bar to have Wallets entry.([#4581](https://github.com/brave/brave-browser/issues/4581))
- [ ] Show BAT by default on main Wallet UI.([#4773](https://github.com/brave/brave-browser/issues/4773))
- [ ] Twitter metadata needs to be generalized for other media publishers.([#4907](https://github.com/brave/brave-browser/issues/4907))
- [ ] Add theme setting as an inline option to onboarding process.([#4992](https://github.com/brave/brave-browser/issues/4992))
- [ ] Dark mode for new tabs page settings.([#5014](https://github.com/brave/brave-browser/issues/5014))
- [ ] Widget menu for new tabs page elements.([#5015](https://github.com/brave/brave-browser/issues/5015))
- [ ] GitHub inline tipping support.([#5040](https://github.com/brave/brave-browser/issues/5040))
- [ ] Hide profile icon until second profile icon is added.([#5091](https://github.com/brave/brave-browser/issues/5091))
- [ ] Brave://wallet should appear as a Brave secure page (if possible).([#5109](https://github.com/brave/brave-browser/issues/5109))
- [ ] No subframe should be present when ethereum-remote-client is installed.([#5110](https://github.com/brave/brave-browser/issues/5110))
- [ ] Map ethereum-remote-client extension path to brave://wallet.([#5117](https://github.com/brave/brave-browser/issues/5117))
- [ ] Auto install ethereum-remote-client and add for usage on first access of brave://wallet.([#5118](https://github.com/brave/brave-browser/issues/5118))
- [ ] Wire SaveRecurringTip through ledger.([#5152](https://github.com/brave/brave-browser/issues/5152))
- [ ] Ad matching for ads that are not categorized.([#5183](https://github.com/brave/brave-browser/issues/5183))
- [ ] Add ability to edit profile in settings.([#5211](https://github.com/brave/brave-browser/issues/5211))
- [ ] Add Profile Manager items to App menu.([#5212](https://github.com/brave/brave-browser/issues/5212))
- [ ] Elements not blocked by selector when tab loaded in background.([#5215](https://github.com/brave/brave-browser/issues/5215))
- [ ] Include a random delay between sending fee and tips for p2p contributions.([#5334](https://github.com/brave/brave-browser/issues/5334))
- [ ] Show All button for download bar not visible with dark theme or on Tor window.([#5350](https://github.com/brave/brave-browser/issues/5350))
- [ ] Sync: Implement exponential re-sends.([#5356](https://github.com/brave/brave-browser/issues/5356))
- [ ] Brave://welcome page background needs to be adjusted for RTL languages.([#5377](https://github.com/brave/brave-browser/issues/5377))
- [ ] Crypto Wallets menu item is present on official builds.([#5383](https://github.com/brave/brave-browser/issues/5383))
- [ ] Logic for shield settings is located in multiple places and none of them are shareable with android.([#5416](https://github.com/brave/brave-browser/issues/5416))
- [ ] Chromium incognito icon used in profile dropdown .([#5432](https://github.com/brave/brave-browser/issues/5432))
- [ ] There is no ads enable alert, ads are enabled by default when upgrading build from 0.62.51 to 0.67.119.([#5434](https://github.com/brave/brave-browser/issues/5434))
- [ ] Exclude toggle crash.([#5457](https://github.com/brave/brave-browser/issues/5457))
- [ ] WebTorrent: Autoplay is sometimes blocked if media metadata is not ready yet.([#5471](https://github.com/brave/brave-browser/issues/5471))
- [ ] Brave ads mobile delivery policy change.([#5507](https://github.com/brave/brave-browser/issues/5507))
- [ ] Remove confirm button from changing theme option in onboarding.([#5509](https://github.com/brave/brave-browser/issues/5509))
- [ ] Welcome Page Scroll Bar Appearing.([#5520](https://github.com/brave/brave-browser/issues/5520))
- [ ] Follow up to there is no ads enable alert, ads are enabled by default when upgrading build from 0.62.51 to 0.67.119.([#5531](https://github.com/brave/brave-browser/issues/5531))
- [ ] Rewards : User Wallets - Verified wallet visual indicator is causing confusion versus verified publishers.([#5543](https://github.com/brave/brave-browser/issues/5543))
- [ ] User feedback loop crash if ads are disabled or region is not supported.([#5659](https://github.com/brave/brave-browser/issues/5659))
- [ ] User wallets - NY loop.([#5799](https://github.com/brave/brave-browser/issues/5799))
- [ ] Replace 5% settlement card with an anon address.([#5808](https://github.com/brave/brave-browser/issues/5808))
- [ ] WebTorrent http server XSS.([#5821](https://github.com/brave/brave-browser/issues/5821))
- [ ] Replace Uphold logo with their updated version. .([#5876](https://github.com/brave/brave-browser/issues/5876))
### Installer
- [ ] Check that installer is close to the size of last release
- [ ] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### Data(Upgrade from previous release)
- [ ] Make sure that data from the last version appears in the new version OK
- [ ] With data from the last version, verify that
- [ ] Bookmarks on the bookmark toolbar and bookmark folders can be opened
- [ ] Cookies are preserved
- [ ] Installed extensions are retained and work correctly
- [ ] Opened tabs can be reloaded
- [ ] Stored passwords are preserved
- [ ] Unpinned tabs can be pinned
- [ ] Sync chain created in previous version is retained
- [ ] Social media blocking buttons changes are retained
### About pages
- [ ] Verify that `chrome://` forwards to `brave://`
- [ ] Verify `brave://adblock` loads adblock page
- [ ] Verify `brave://newtab` loads a new tab
- [ ] Verify `brave://rewards` loads Brave rewards page
- [ ] Verify `brave://settings` loads Brave settings page
- [ ] Verify `brave://version` correctly shows Brave version and Chromium version
- [ ] Verify `brave://welcome` loads the welcome page
### Import tests
- [ ] Verify that you can import bookmarks, cookies, history, passwords from Google Chrome
- [ ] Verify that you can import bookmarks, cookies, history, passwords, autofill and search engines from Firefox
- [ ] Verify that you can import bookmarks from Edge
- [ ] Verify importing `Bookmark HTML` file retains the folder structure on a clean profile
### Context menus
- [ ] Verify `Block element via selector` removes a CSS element from page without reloading
- [ ] Verify `Clear all CSS rules for this sites` removes the blocked element after page reload
- [ ] Verify `Clear all CSS rules for all sites` removes the rule and loads all elements for all pages
## Extensions/Plugins tests
- [ ] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension)
- [ ] Verify one item from Google Update server is installable (Example: Extensions from CWS)
- [ ] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled
- [ ] Verify older version of an extension gets updated to new version via Google server
- [ ] Verify older version of an extension gets updated to new version via Brave server
- [ ] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent
- [ ] Use an old profile to verify extension updates work correctly.
### CWS
- [ ] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!`but still allows to install the extension
- [ ] Verify installing LastPass from CWS doesn't show any warning message
- [ ] Verify installing an extension that is not vetted by Brave gets blocked
### Flash tests
- [ ] Test that Flash gets blocked by default when installed
- [ ] Test that once you allow Flash, it turns into a click to play area, and loads flash when allowed
### PDF
- [ ] Test that you can print a PDF
- [ ] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf
- [ ] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf
### Widevine
- [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine
### Geolocation
- [ ] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation shows correct location
- [ ] Check that https://browserleaks.com/geo works and shows correct location
- [ ] Check that https://html5demos.com/geo/ works but doesn't require an accurate location
### Crash Reporting
- [ ] Check that loading `brave://crash` causes the new tab to crash
- [ ] Check that `brave://crashes` lists all the crashes and includes both Crash Report ID & Local Crash ID
- [ ] Verify the crash ID matches the report on brave stats
### Performance test
_Each start should take less than 7 seconds_
- [ ] Enable only rewards
- [ ] Only import a large set of bookmarks
- [ ] Combine rewards, and a large set of bookmarks
### Bravery settings
- [ ] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [ ] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/
- [ ] Verify that toggling `Ads and trackers blocked` works as expected
- [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [ ] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly
- [ ] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
### Fingerprint Tests
- [ ] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00`
- [ ] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on
- [ ] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/
- [ ] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on
### Content tests
- [ ] Open a page with an input control and type some misspellings on a textbox, make sure they are underlined
- [ ] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text
- [ ] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run)
### Brave Ads
- [ ] Verify when you enable Rewards from panel or brave://rewards, Ads are enabled by default
- [ ] Verify Ads UI (panel, settings, etc) shows when in a region with Ads support
- [ ] Verify Ads UI (panel, settings, etc) does not show when in a region without Ads support. Verify the Ads panel does show the 'Sorry! Ads are not yet available in your region.' message.
- [ ] Verify when the system language is English, the Browser language is French, and you are in one of the supported regions, Ad notifications are still served to you.
- [ ] Verify you are served Ad notifications when Ads are enabled
- [ ] Windows only: Verify when `Focus Assist` is set to alarms only, Ads are not served to the user but are available in the `Notification Center`
- [ ] Verify when Ads are toggled off, there are no Ad messages in the logs
- [ ] Verify when Rewards are toggled off (but Ads were not explicitly toggled off), there are no Ads logs recorded
- [ ] Verify view/click/dismiss/landed ad notifications show in confirmations.json
- [ ] Verify pages you browse to are being classified in the logs
- [ ] Verify tokens are redeemed by viewing the logs (you can use --rewards=debug=true to shorten redemption time)
- [ ] Verify Ad is not shown if a tab is playing media and is only shown after it stops playing
- [ ] Upgrade Cases
- [ ] Verify when updating from the previous version Ads info in the panel is not lost
- [ ] Update to latest version from 0.62.51 (or a version without Ads available). Verify you are notified to try ads via a BAT logo notification but Ads are not automatically turned on for you.
- [ ] Update to latest version from some Ads enabled versions (0.63.x, 0.64.x). Verify Ads status (on/off) is the same as it was prior to update.
### Rewards
- [ ] Verify wallet is auto created after enabling rewards
- [ ] Verify account balance shows correct BAT and USD value
- [ ] Verify you are able to restore a wallet
- [ ] Verify wallet address matches the QR code that is generated under `Add funds`
- [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel
- [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time
- [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately
- [ ] Verify monthly budget shows correct BAT and USD value
- [ ] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites
- [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel
- [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %)
- [ ] Verify when you click on `Send a tip`, the custom tip banner displays
- [ ] Verify you are able to make one-time tip and they display in tips panel
- [ ] Verify you are able to make recurring tip and they display in tips panel
- [ ] Verify you can tip a verified publisher
- [ ] Verify you can tip a verified YouTube creator
- [ ] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator
- [ ] Verify tip panel shows a message about unverified publisher
- [ ] Verify BR panel shows message about an unverified publisher
- [ ] Verify you are able to perform a contribution
- [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators
- [ ] Verify that disabling Rewards and enabling it again does not lose state
- [ ] Verify that disabling auto-contribute and enabling it again does not lose state
- [ ] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings
- [ ] Upgrade from older version
- [ ] Verify the wallet balance is retained and wallet backup code isn't corrupted
- [ ] Verify auto-contribute list is not lost after upgrade
- [ ] Verify tips list is not lost after upgrade
- [ ] Verify wallet panel transactions list is not lost after upgrade
### Social-media blocking settings
- [ ] Verify individual `Social media blocking` buttons works as intended when enabled/disabled by visiting https://fmarier.github.io/brave-testing/social-widgets.html
### Sync
- [ ] Verify you are able to create a sync chain and add a mobile/computer to the chain
- [ ] Verify you are able to join an existing sync chain using code words
- [ ] Verify the device name is shown properly when sync chain is created
- [ ] Verify you are able to add a new mobile device to the chain via QR code/code words
- [ ] Verify newly created bookmarks get sync'd to all devices on the sync chain
- [ ] Verify existing bookmarks on current profile gets sync'd to all devices on the sync chain
- [ ] Verify folder structure is retained after sync completes
- [ ] Verify bookmarks don't duplicate when sync'd from other devices
- [ ] Verify removing bookmark from device gets sync'd to all devices on the sync chain
- [ ] Verify adding/removing a bookmark in offline mode gets sync'd to all devices on the sync chain when device comes online
- [ ] With only two device in chain, verify removing the other device resets the sync on b-c as well
### Tor Tabs
- [ ] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Click `New Tor Identity for this site` in app menu, ensure the exit node IP address changes after page is reloaded
- [ ] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown
- [ ] Visit http://pdf995.com in a Tor window, should warn about connecting to HTTP site
- [ ] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown
- [ ] Visit https://diafygi.github.io/webrtc-ips/ in a Tor window with block all fingerprinting, ensure WebRTC is blocked and no IP is shown
- [ ] Visit https://diafygi.github.io/webrtc-ips/ in a Tor window, disable shields, ensure WebRTC is blocked and no IP is shown
- [ ] Verify Flash doesn't work on Tor window even if it is enabled in `brave://settings/content/flash`
- [ ] Verify Torrent viewer doesn't load in a Tor window and warns when trying to load a torrent/magnet link in a Tor window
- [ ] Verify Google Widevine doesn't load in Tor window and doesn't prompt Google Windevine notification in URL bar
- [ ] Ensure you are able to download a file in Tor window. Verify all Download/Cancel, Download/Retry and Download works in Tor window
- [ ] Disconnect network and open a Tor window, should show modal to retry connection or relaunch Tor window to connect
### Cookie and Cache
- [ ] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the evercookie site does not remember the old evercookie value
### Session storage
- [ ] Temporarily move away your browser profile and test that a new profile is created when browser is launched
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [ ] Test that windows and tabs restore when closed, including active tab
- [ ] Ensure that the tabs in the above session are being lazy loaded when the session is restored
## Update tests
- [ ] Verify visiting `brave://settings/help` triggers update check
- [ ] Verify once update is downloaded, prompts to `Relaunch` to install update
## Chromium upgrade tests
- [ ] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers
#### Adblock
- [ ] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/`
#### Components
- [ ] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
- [ ] Restart the browser, load brave://components, wait for 8 mins and verify that no component shows any errors
- [ ] Verify Brave Local Data Updater, Brave Ad Block Updater, Brave Tor Client Updater (Mac), PDF Viewer (PDF.js), Brave HTTPS Everywhere Updater have non-zero version numbers
### Keyboard Shortcuts (WIP)
#### Rewards Media (To be verified on YouTube and Twitch) (WIP)
| 1.0 | Manual test run on Linux for 0.69.x - Release - ## Per release specialty tests
- [ ] Chromecast support.([#209](https://github.com/brave/brave-browser/issues/209))
- [ ] Add Dapp detection.([#718](https://github.com/brave/brave-browser/issues/718))
- [ ] Scale cover image height on tipping banner.([#2015](https://github.com/brave/brave-browser/issues/2015))
- [ ] Migrate sync code to Syncable Service.([#2754](https://github.com/brave/brave-browser/issues/2754))
- [ ] Cosmetic filter !important.([#3041](https://github.com/brave/brave-browser/issues/3041))
- [ ] Warning message not being displayed, when click on `Confirm Sync code` without any code added.([#3190](https://github.com/brave/brave-browser/issues/3190))
- [ ] Welcome screen: support dark mode .([#3911](https://github.com/brave/brave-browser/issues/3911))
- [ ] Implement user feedback loop.([#4047](https://github.com/brave/brave-browser/issues/4047))
- [ ] Blocked ads sometimes leave additional white space on page.([#4338](https://github.com/brave/brave-browser/issues/4338))
- [ ] Custom adblock rules should allow cosmetic filters .([#4348](https://github.com/brave/brave-browser/issues/4348))
- [ ] Add built-in Ethereum remote client / wallet / Dapp browser.([#4494](https://github.com/brave/brave-browser/issues/4494))
- [ ] Add a menu item for triggering WebUI for the Ethereum wallet.([#4579](https://github.com/brave/brave-browser/issues/4579))
- [ ] Settings bar to have Wallets entry.([#4581](https://github.com/brave/brave-browser/issues/4581))
- [ ] Show BAT by default on main Wallet UI.([#4773](https://github.com/brave/brave-browser/issues/4773))
- [ ] Twitter metadata needs to be generalized for other media publishers.([#4907](https://github.com/brave/brave-browser/issues/4907))
- [ ] Add theme setting as an inline option to onboarding process.([#4992](https://github.com/brave/brave-browser/issues/4992))
- [ ] Dark mode for new tabs page settings.([#5014](https://github.com/brave/brave-browser/issues/5014))
- [ ] Widget menu for new tabs page elements.([#5015](https://github.com/brave/brave-browser/issues/5015))
- [ ] GitHub inline tipping support.([#5040](https://github.com/brave/brave-browser/issues/5040))
- [ ] Hide profile icon until second profile icon is added.([#5091](https://github.com/brave/brave-browser/issues/5091))
- [ ] Brave://wallet should appear as a Brave secure page (if possible).([#5109](https://github.com/brave/brave-browser/issues/5109))
- [ ] No subframe should be present when ethereum-remote-client is installed.([#5110](https://github.com/brave/brave-browser/issues/5110))
- [ ] Map ethereum-remote-client extension path to brave://wallet.([#5117](https://github.com/brave/brave-browser/issues/5117))
- [ ] Auto install ethereum-remote-client and add for usage on first access of brave://wallet.([#5118](https://github.com/brave/brave-browser/issues/5118))
- [ ] Wire SaveRecurringTip through ledger.([#5152](https://github.com/brave/brave-browser/issues/5152))
- [ ] Ad matching for ads that are not categorized.([#5183](https://github.com/brave/brave-browser/issues/5183))
- [ ] Add ability to edit profile in settings.([#5211](https://github.com/brave/brave-browser/issues/5211))
- [ ] Add Profile Manager items to App menu.([#5212](https://github.com/brave/brave-browser/issues/5212))
- [ ] Elements not blocked by selector when tab loaded in background.([#5215](https://github.com/brave/brave-browser/issues/5215))
- [ ] Include a random delay between sending fee and tips for p2p contributions.([#5334](https://github.com/brave/brave-browser/issues/5334))
- [ ] Show All button for download bar not visible with dark theme or on Tor window.([#5350](https://github.com/brave/brave-browser/issues/5350))
- [ ] Sync: Implement exponential re-sends.([#5356](https://github.com/brave/brave-browser/issues/5356))
- [ ] Brave://welcome page background needs to be adjusted for RTL languages.([#5377](https://github.com/brave/brave-browser/issues/5377))
- [ ] Crypto Wallets menu item is present on official builds.([#5383](https://github.com/brave/brave-browser/issues/5383))
- [ ] Logic for shield settings is located in multiple places and none of them are shareable with android.([#5416](https://github.com/brave/brave-browser/issues/5416))
- [ ] Chromium incognito icon used in profile dropdown .([#5432](https://github.com/brave/brave-browser/issues/5432))
- [ ] There is no ads enable alert, ads are enabled by default when upgrading build from 0.62.51 to 0.67.119.([#5434](https://github.com/brave/brave-browser/issues/5434))
- [ ] Exclude toggle crash.([#5457](https://github.com/brave/brave-browser/issues/5457))
- [ ] WebTorrent: Autoplay is sometimes blocked if media metadata is not ready yet.([#5471](https://github.com/brave/brave-browser/issues/5471))
- [ ] Brave ads mobile delivery policy change.([#5507](https://github.com/brave/brave-browser/issues/5507))
- [ ] Remove confirm button from changing theme option in onboarding.([#5509](https://github.com/brave/brave-browser/issues/5509))
- [ ] Welcome Page Scroll Bar Appearing.([#5520](https://github.com/brave/brave-browser/issues/5520))
- [ ] Follow up to there is no ads enable alert, ads are enabled by default when upgrading build from 0.62.51 to 0.67.119.([#5531](https://github.com/brave/brave-browser/issues/5531))
- [ ] Rewards : User Wallets - Verified wallet visual indicator is causing confusion versus verified publishers.([#5543](https://github.com/brave/brave-browser/issues/5543))
- [ ] User feedback loop crash if ads are disabled or region is not supported.([#5659](https://github.com/brave/brave-browser/issues/5659))
- [ ] User wallets - NY loop.([#5799](https://github.com/brave/brave-browser/issues/5799))
- [ ] Replace 5% settlement card with an anon address.([#5808](https://github.com/brave/brave-browser/issues/5808))
- [ ] WebTorrent http server XSS.([#5821](https://github.com/brave/brave-browser/issues/5821))
- [ ] Replace Uphold logo with their updated version. .([#5876](https://github.com/brave/brave-browser/issues/5876))
### Installer
- [ ] Check that installer is close to the size of last release
- [ ] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### Data(Upgrade from previous release)
- [ ] Make sure that data from the last version appears in the new version OK
- [ ] With data from the last version, verify that
- [ ] Bookmarks on the bookmark toolbar and bookmark folders can be opened
- [ ] Cookies are preserved
- [ ] Installed extensions are retained and work correctly
- [ ] Opened tabs can be reloaded
- [ ] Stored passwords are preserved
- [ ] Unpinned tabs can be pinned
- [ ] Sync chain created in previous version is retained
- [ ] Social media blocking buttons changes are retained
### About pages
- [ ] Verify that `chrome://` forwards to `brave://`
- [ ] Verify `brave://adblock` loads adblock page
- [ ] Verify `brave://newtab` loads a new tab
- [ ] Verify `brave://rewards` loads Brave rewards page
- [ ] Verify `brave://settings` loads Brave settings page
- [ ] Verify `brave://version` correctly shows Brave version and Chromium version
- [ ] Verify `brave://welcome` loads the welcome page
### Import tests
- [ ] Verify that you can import bookmarks, cookies, history, passwords from Google Chrome
- [ ] Verify that you can import bookmarks, cookies, history, passwords, autofill and search engines from Firefox
- [ ] Verify that you can import bookmarks from Edge
- [ ] Verify importing `Bookmark HTML` file retains the folder structure on a clean profile
### Context menus
- [ ] Verify `Block element via selector` removes a CSS element from page without reloading
- [ ] Verify `Clear all CSS rules for this sites` removes the blocked element after page reload
- [ ] Verify `Clear all CSS rules for all sites` removes the rule and loads all elements for all pages
## Extensions/Plugins tests
- [ ] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension)
- [ ] Verify one item from Google Update server is installable (Example: Extensions from CWS)
- [ ] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled
- [ ] Verify older version of an extension gets updated to new version via Google server
- [ ] Verify older version of an extension gets updated to new version via Brave server
- [ ] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent
- [ ] Use an old profile to verify extension updates work correctly.
### CWS
- [ ] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!`but still allows to install the extension
- [ ] Verify installing LastPass from CWS doesn't show any warning message
- [ ] Verify installing an extension that is not vetted by Brave gets blocked
### Flash tests
- [ ] Test that Flash gets blocked by default when installed
- [ ] Test that once you allow Flash, it turns into a click to play area, and loads flash when allowed
### PDF
- [ ] Test that you can print a PDF
- [ ] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf
- [ ] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf
### Widevine
- [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine
### Geolocation
- [ ] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation shows correct location
- [ ] Check that https://browserleaks.com/geo works and shows correct location
- [ ] Check that https://html5demos.com/geo/ works but doesn't require an accurate location
### Crash Reporting
- [ ] Check that loading `brave://crash` causes the new tab to crash
- [ ] Check that `brave://crashes` lists all the crashes and includes both Crash Report ID & Local Crash ID
- [ ] Verify the crash ID matches the report on brave stats
### Performance test
_Each start should take less than 7 seconds_
- [ ] Enable only rewards
- [ ] Only import a large set of bookmarks
- [ ] Combine rewards, and a large set of bookmarks
### Bravery settings
- [ ] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [ ] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/
- [ ] Verify that toggling `Ads and trackers blocked` works as expected
- [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [ ] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly
- [ ] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
### Fingerprint Tests
- [ ] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00`
- [ ] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on
- [ ] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/
- [ ] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on
### Content tests
- [ ] Open a page with an input control and type some misspellings on a textbox, make sure they are underlined
- [ ] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text
- [ ] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run)
### Brave Ads
- [ ] Verify when you enable Rewards from panel or brave://rewards, Ads are enabled by default
- [ ] Verify Ads UI (panel, settings, etc) shows when in a region with Ads support
- [ ] Verify Ads UI (panel, settings, etc) does not show when in a region without Ads support. Verify the Ads panel does show the 'Sorry! Ads are not yet available in your region.' message.
- [ ] Verify when the system language is English, the Browser language is French, and you are in one of the supported regions, Ad notifications are still served to you.
- [ ] Verify you are served Ad notifications when Ads are enabled
- [ ] Windows only: Verify when `Focus Assist` is set to alarms only, Ads are not served to the user but are available in the `Notification Center`
- [ ] Verify when Ads are toggled off, there are no Ad messages in the logs
- [ ] Verify when Rewards are toggled off (but Ads were not explicitly toggled off), there are no Ads logs recorded
- [ ] Verify view/click/dismiss/landed ad notifications show in confirmations.json
- [ ] Verify pages you browse to are being classified in the logs
- [ ] Verify tokens are redeemed by viewing the logs (you can use --rewards=debug=true to shorten redemption time)
- [ ] Verify Ad is not shown if a tab is playing media and is only shown after it stops playing
- [ ] Upgrade Cases
- [ ] Verify when updating from the previous version Ads info in the panel is not lost
- [ ] Update to latest version from 0.62.51 (or a version without Ads available). Verify you are notified to try ads via a BAT logo notification but Ads are not automatically turned on for you.
- [ ] Update to latest version from some Ads enabled versions (0.63.x, 0.64.x). Verify Ads status (on/off) is the same as it was prior to update.
### Rewards
- [ ] Verify wallet is auto created after enabling rewards
- [ ] Verify account balance shows correct BAT and USD value
- [ ] Verify you are able to restore a wallet
- [ ] Verify wallet address matches the QR code that is generated under `Add funds`
- [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel
- [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time
- [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately
- [ ] Verify monthly budget shows correct BAT and USD value
- [ ] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites
- [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel
- [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %)
- [ ] Verify when you click on `Send a tip`, the custom tip banner displays
- [ ] Verify you are able to make one-time tip and they display in tips panel
- [ ] Verify you are able to make recurring tip and they display in tips panel
- [ ] Verify you can tip a verified publisher
- [ ] Verify you can tip a verified YouTube creator
- [ ] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator
- [ ] Verify tip panel shows a message about unverified publisher
- [ ] Verify BR panel shows message about an unverified publisher
- [ ] Verify you are able to perform a contribution
- [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators
- [ ] Verify that disabling Rewards and enabling it again does not lose state
- [ ] Verify that disabling auto-contribute and enabling it again does not lose state
- [ ] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings
- [ ] Upgrade from older version
- [ ] Verify the wallet balance is retained and wallet backup code isn't corrupted
- [ ] Verify auto-contribute list is not lost after upgrade
- [ ] Verify tips list is not lost after upgrade
- [ ] Verify wallet panel transactions list is not lost after upgrade
### Social-media blocking settings
- [ ] Verify individual `Social media blocking` buttons works as intended when enabled/disabled by visiting https://fmarier.github.io/brave-testing/social-widgets.html
### Sync
- [ ] Verify you are able to create a sync chain and add a mobile/computer to the chain
- [ ] Verify you are able to join an existing sync chain using code words
- [ ] Verify the device name is shown properly when sync chain is created
- [ ] Verify you are able to add a new mobile device to the chain via QR code/code words
- [ ] Verify newly created bookmarks get sync'd to all devices on the sync chain
- [ ] Verify existing bookmarks on current profile gets sync'd to all devices on the sync chain
- [ ] Verify folder structure is retained after sync completes
- [ ] Verify bookmarks don't duplicate when sync'd from other devices
- [ ] Verify removing bookmark from device gets sync'd to all devices on the sync chain
- [ ] Verify adding/removing a bookmark in offline mode gets sync'd to all devices on the sync chain when device comes online
- [ ] With only two device in chain, verify removing the other device resets the sync on b-c as well
### Tor Tabs
- [ ] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Click `New Tor Identity for this site` in app menu, ensure the exit node IP address changes after page is reloaded
- [ ] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown
- [ ] Visit http://pdf995.com in a Tor window, should warn about connecting to HTTP site
- [ ] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown
- [ ] Visit https://diafygi.github.io/webrtc-ips/ in a Tor window with block all fingerprinting, ensure WebRTC is blocked and no IP is shown
- [ ] Visit https://diafygi.github.io/webrtc-ips/ in a Tor window, disable shields, ensure WebRTC is blocked and no IP is shown
- [ ] Verify Flash doesn't work on Tor window even if it is enabled in `brave://settings/content/flash`
- [ ] Verify Torrent viewer doesn't load in a Tor window and warns when trying to load a torrent/magnet link in a Tor window
- [ ] Verify Google Widevine doesn't load in Tor window and doesn't prompt Google Windevine notification in URL bar
- [ ] Ensure you are able to download a file in Tor window. Verify all Download/Cancel, Download/Retry and Download works in Tor window
- [ ] Disconnect network and open a Tor window, should show modal to retry connection or relaunch Tor window to connect
### Cookie and Cache
- [ ] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the evercookie site does not remember the old evercookie value
### Session storage
- [ ] Temporarily move away your browser profile and test that a new profile is created when browser is launched
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [ ] Test that windows and tabs restore when closed, including active tab
- [ ] Ensure that the tabs in the above session are being lazy loaded when the session is restored
## Update tests
- [ ] Verify visiting `brave://settings/help` triggers update check
- [ ] Verify once update is downloaded, prompts to `Relaunch` to install update
## Chromium upgrade tests
- [ ] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers
#### Adblock
- [ ] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/`
#### Components
- [ ] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
- [ ] Restart the browser, load brave://components, wait for 8 mins and verify that no component shows any errors
- [ ] Verify Brave Local Data Updater, Brave Ad Block Updater, Brave Tor Client Updater (Mac), PDF Viewer (PDF.js), Brave HTTPS Everywhere Updater have non-zero version numbers
### Keyboard Shortcuts (WIP)
#### Rewards Media (To be verified on YouTube and Twitch) (WIP)
| non_priority | manual test run on linux for x release per release specialty tests chromecast support add dapp detection scale cover image height on tipping banner migrate sync code to syncable service cosmetic filter important warning message not being displayed when click on confirm sync code without any code added welcome screen support dark mode implement user feedback loop blocked ads sometimes leave additional white space on page custom adblock rules should allow cosmetic filters add built in ethereum remote client wallet dapp browser add a menu item for triggering webui for the ethereum wallet settings bar to have wallets entry show bat by default on main wallet ui twitter metadata needs to be generalized for other media publishers add theme setting as an inline option to onboarding process dark mode for new tabs page settings widget menu for new tabs page elements github inline tipping support hide profile icon until second profile icon is added brave wallet should appear as a brave secure page if possible no subframe should be present when ethereum remote client is installed map ethereum remote client extension path to brave wallet auto install ethereum remote client and add for usage on first access of brave wallet wire saverecurringtip through ledger ad matching for ads that are not categorized add ability to edit profile in settings add profile manager items to app menu elements not blocked by selector when tab loaded in background include a random delay between sending fee and tips for contributions show all button for download bar not visible with dark theme or on tor window sync implement exponential re sends brave welcome page background needs to be adjusted for rtl languages crypto wallets menu item is present on official builds logic for shield settings is located in multiple places and none of them are shareable with android chromium incognito icon used in profile dropdown there is no ads enable alert ads are enabled by default when upgrading build from to exclude toggle crash webtorrent autoplay is sometimes blocked if media metadata is not ready yet brave ads mobile delivery policy change remove confirm button from changing theme option in onboarding welcome page scroll bar appearing follow up to there is no ads enable alert ads are enabled by default when upgrading build from to rewards user wallets verified wallet visual indicator is causing confusion versus verified publishers user feedback loop crash if ads are disabled or region is not supported user wallets ny loop replace settlement card with an anon address webtorrent http server xss replace uphold logo with their updated version installer check that installer is close to the size of last release check signature if os run spctl assess verbose applications brave browser beta app and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window data upgrade from previous release make sure that data from the last version appears in the new version ok with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved unpinned tabs can be pinned sync chain created in previous version is retained social media blocking buttons changes are retained about pages verify that chrome forwards to brave verify brave adblock loads adblock page verify brave newtab loads a new tab verify brave rewards loads brave rewards page verify brave settings loads brave settings page verify brave version correctly shows brave version and chromium version verify brave welcome loads the welcome page import tests verify that you can import bookmarks cookies history passwords from google chrome verify that you can import bookmarks cookies history passwords autofill and search engines from firefox verify that you can import bookmarks from edge verify importing bookmark html file retains the folder structure on a clean profile context menus verify block element via selector removes a css element from page without reloading verify clear all css rules for this sites removes the blocked element after page reload verify clear all css rules for all sites removes the rule and loads all elements for all pages extensions plugins tests verify one item from brave update server is installable example ad block dat file on fresh extension verify one item from google update server is installable example extensions from cws verify pdfjs torrent viewer extensions are installed automatically on fresh profile and cannot be disabled verify older version of an extension gets updated to new version via google server verify older version of an extension gets updated to new version via brave server verify magnet links and torrent files loads torrent viewer page and able to download torrent use an old profile to verify extension updates work correctly cws verify installing abp from cws shows warning message not a recommended brave extension but still allows to install the extension verify installing lastpass from cws doesn t show any warning message verify installing an extension that is not vetted by brave gets blocked flash tests test that flash gets blocked by default when installed test that once you allow flash it turns into a click to play area and loads flash when allowed pdf test that you can print a pdf test that pdf is loaded over https at test that pdf is loaded over http at widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine geolocation check that shows correct location check that works and shows correct location check that works but doesn t require an accurate location crash reporting check that loading brave crash causes the new tab to crash check that brave crashes lists all the crashes and includes both crash report id local crash id verify the crash id matches the report on brave stats performance test each start should take less than seconds enable only rewards only import a large set of bookmarks combine rewards and a large set of bookmarks bravery settings verify that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to verify that toggling ads and trackers blocked works as expected visit verify that safe browsing via our proxy works for all the listed items visit and then turn on script blocking page should not load allow it from the script blocking ui in the url bar and it should load the page correctly test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked fingerprint tests visit ensure blocked items are listed in shields result window should show got canvas fingerprint and got webgl fingerprint test that audio fingerprint is blocked at only when block all fingerprinting protection is on test that brave browser isn t detected on test that doesn t leak ip address when block all fingerprinting protection is on content tests open a page with an input control and type some misspellings on a textbox make sure they are underlined make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text test that shows up as grey not red no mixed content scripts are run brave ads verify when you enable rewards from panel or brave rewards ads are enabled by default verify ads ui panel settings etc shows when in a region with ads support verify ads ui panel settings etc does not show when in a region without ads support verify the ads panel does show the sorry ads are not yet available in your region message verify when the system language is english the browser language is french and you are in one of the supported regions ad notifications are still served to you verify you are served ad notifications when ads are enabled windows only verify when focus assist is set to alarms only ads are not served to the user but are available in the notification center verify when ads are toggled off there are no ad messages in the logs verify when rewards are toggled off but ads were not explicitly toggled off there are no ads logs recorded verify view click dismiss landed ad notifications show in confirmations json verify pages you browse to are being classified in the logs verify tokens are redeemed by viewing the logs you can use rewards debug true to shorten redemption time verify ad is not shown if a tab is playing media and is only shown after it stops playing upgrade cases verify when updating from the previous version ads info in the panel is not lost update to latest version from or a version without ads available verify you are notified to try ads via a bat logo notification but ads are not automatically turned on for you update to latest version from some ads enabled versions x x verify ads status on off is the same as it was prior to update rewards verify wallet is auto created after enabling rewards verify account balance shows correct bat and usd value verify you are able to restore a wallet verify wallet address matches the qr code that is generated under add funds verify actions taken claiming grant tipping auto contribute display in wallet panel verify adding funds via any of the currencies flows into wallet after specified amount of time verify adding funds to an existing wallet with amount adjusts the bat value appropriately verify monthly budget shows correct bat and usd value verify you are able to exclude a publisher from the auto contribute table by clicking on the x in auto contribute table and popup list of sites verify you are able to exclude a publisher by using the toggle on the rewards panel verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify when you click on send a tip the custom tip banner displays verify you are able to make one time tip and they display in tips panel verify you are able to make recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify tip panel shows a verified checkmark for a verified publisher verified youtube creator verify tip panel shows a message about unverified publisher verify br panel shows message about an unverified publisher verify you are able to perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators verify that disabling rewards and enabling it again does not lose state verify that disabling auto contribute and enabling it again does not lose state adjust min visit time in settings visit some sites and youtube channels to verify they are added to the table after the specified settings upgrade from older version verify the wallet balance is retained and wallet backup code isn t corrupted verify auto contribute list is not lost after upgrade verify tips list is not lost after upgrade verify wallet panel transactions list is not lost after upgrade social media blocking settings verify individual social media blocking buttons works as intended when enabled disabled by visiting sync verify you are able to create a sync chain and add a mobile computer to the chain verify you are able to join an existing sync chain using code words verify the device name is shown properly when sync chain is created verify you are able to add a new mobile device to the chain via qr code code words verify newly created bookmarks get sync d to all devices on the sync chain verify existing bookmarks on current profile gets sync d to all devices on the sync chain verify folder structure is retained after sync completes verify bookmarks don t duplicate when sync d from other devices verify removing bookmark from device gets sync d to all devices on the sync chain verify adding removing a bookmark in offline mode gets sync d to all devices on the sync chain when device comes online with only two device in chain verify removing the other device resets the sync on b c as well tor tabs visit in a tor window ensure its shows success message for using a tor exit node visit in a tor window note down exit node ip address do a hard refresh ctrl shift r cmd shift r ensure exit ip changes after page reloads visit in a tor window note down exit node ip address click new tor identity for this site in app menu ensure the exit node ip address changes after page is reloaded visit in a tor window ensure login page is shown visit in a tor window should warn about connecting to http site visit in a tor window ensure location isn t shown visit in a tor window with block all fingerprinting ensure webrtc is blocked and no ip is shown visit in a tor window disable shields ensure webrtc is blocked and no ip is shown verify flash doesn t work on tor window even if it is enabled in brave settings content flash verify torrent viewer doesn t load in a tor window and warns when trying to load a torrent magnet link in a tor window verify google widevine doesn t load in tor window and doesn t prompt google windevine notification in url bar ensure you are able to download a file in tor window verify all download cancel download retry and download works in tor window disconnect network and open a tor window should show modal to retry connection or relaunch tor window to connect cookie and cache go to and set an evercookie check that going to prefs clearing site data and cache and going back to the evercookie site does not remember the old evercookie value session storage temporarily move away your browser profile and test that a new profile is created when browser is launched macos library application support bravesoftware windows userprofile appdata local bravesoftware linux ubuntu config bravesoftware test that windows and tabs restore when closed including active tab ensure that the tabs in the above session are being lazy loaded when the session is restored update tests verify visiting brave settings help triggers update check verify once update is downloaded prompts to relaunch to install update chromium upgrade tests verify brave gpu on brave and chrome gpu on chrome are similar for the same chromium version on both browsers adblock verify referrer blocking works properly for tld visit and verify adblock works properly similar to components delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components restart the browser load brave components wait for mins and verify that no component shows any errors verify brave local data updater brave ad block updater brave tor client updater mac pdf viewer pdf js brave https everywhere updater have non zero version numbers keyboard shortcuts wip rewards media to be verified on youtube and twitch wip | 0 |
31,981 | 26,322,440,957 | IssuesEvent | 2023-01-10 01:38:33 | crossplane/docs | https://api.github.com/repos/crossplane/docs | closed | Add deploy preview banner | enhancement P2 infrastructure | We should generate a special banner (or similar header) that provides a link back to the PR for [deploy previews](https://docs.netlify.com/site-deploys/deploy-previews/). | 1.0 | Add deploy preview banner - We should generate a special banner (or similar header) that provides a link back to the PR for [deploy previews](https://docs.netlify.com/site-deploys/deploy-previews/). | non_priority | add deploy preview banner we should generate a special banner or similar header that provides a link back to the pr for | 0 |
42,036 | 12,867,875,851 | IssuesEvent | 2020-07-10 07:48:49 | benchmarkdebricked/sentry | https://api.github.com/repos/benchmarkdebricked/sentry | opened | CVE-2020-10379 (High) detected in Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2020-10379 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/43/5a/904f2cc20ef9f9ba05f9ff1fb3dfadb1e6923e3bf6f8c8363d5dc3a179ab/Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/43/5a/904f2cc20ef9f9ba05f9ff1fb3dfadb1e6923e3bf6f8c8363d5dc3a179ab/Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/sentry</p>
<p>Path to vulnerable library: /sentry</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/sentry/commit/3f63a21af8bb1c66de6f859ad79ababfd2c21a46">3f63a21af8bb1c66de6f859ad79ababfd2c21a46</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Pillow before 7.1.0, there are two Buffer Overflows in libImaging/TiffDecode.c.
<p>Publish Date: 2020-06-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10379>CVE-2020-10379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8">https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8</a></p>
<p>Release Date: 2020-06-25</p>
<p>Fix Resolution: 7.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-10379 (High) detected in Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2020-10379 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/43/5a/904f2cc20ef9f9ba05f9ff1fb3dfadb1e6923e3bf6f8c8363d5dc3a179ab/Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/43/5a/904f2cc20ef9f9ba05f9ff1fb3dfadb1e6923e3bf6f8c8363d5dc3a179ab/Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/sentry</p>
<p>Path to vulnerable library: /sentry</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-4.2.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/sentry/commit/3f63a21af8bb1c66de6f859ad79ababfd2c21a46">3f63a21af8bb1c66de6f859ad79ababfd2c21a46</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Pillow before 7.1.0, there are two Buffer Overflows in libImaging/TiffDecode.c.
<p>Publish Date: 2020-06-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10379>CVE-2020-10379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8">https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8</a></p>
<p>Release Date: 2020-06-25</p>
<p>Fix Resolution: 7.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file tmp ws scm sentry path to vulnerable library sentry dependency hierarchy x pillow whl vulnerable library found in head commit a href vulnerability details in pillow before there are two buffer overflows in libimaging tiffdecode c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
224,563 | 17,755,804,082 | IssuesEvent | 2021-08-28 18:33:00 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | ccl/benchccl/rttanalysisccl: BenchmarkAlterPrimaryRegion failed | C-test-failure O-robot branch-master | ccl/benchccl/rttanalysisccl.BenchmarkAlterPrimaryRegion [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=artifacts#/) on master @ [93d95b97e53b11b9640a698535c563ca3ed5ba15](https://github.com/cockroachdb/cockroach/commits/93d95b97e53b11b9640a698535c563ca3ed5ba15):
```
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region651687629
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterTableLocality/alter_from_global_to_rbr-16 1 3283756600 ns/op 22.00 roundtrips
BenchmarkAlterTableLocality/alter_from_regional_by_table_to_rbr
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr712218057
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_drop_region564512728
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterSurvivalGoals/alter_populated_database_from_zone_to_region-16 1 359332916 ns/op 41.00 roundtrips
BenchmarkAlterSurvivalGoals/alter_populated_database_from_region_to_zone
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone141752125
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region085553020
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr356440840
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone849513004
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr215803079
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
rtt_analysis_bench.go:131: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region053354155
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone938059163
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr701030222
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region037082402
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone648673362
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterRegions/alter_populated_database_drop_region-16 1 15275801616 ns/op 21.00 roundtrips
BenchmarkAlterRegions/alter_populated_database_add_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_add_region041894743
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region-16 1 325849061 ns/op 21.00 roundtrips
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=BenchmarkAlterPrimaryRegion PKG=./pkg/ccl/benchccl/rttanalysisccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience pawalt
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*BenchmarkAlterPrimaryRegion.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 1.0 | ccl/benchccl/rttanalysisccl: BenchmarkAlterPrimaryRegion failed - ccl/benchccl/rttanalysisccl.BenchmarkAlterPrimaryRegion [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=artifacts#/) on master @ [93d95b97e53b11b9640a698535c563ca3ed5ba15](https://github.com/cockroachdb/cockroach/commits/93d95b97e53b11b9640a698535c563ca3ed5ba15):
```
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region651687629
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterTableLocality/alter_from_global_to_rbr-16 1 3283756600 ns/op 22.00 roundtrips
BenchmarkAlterTableLocality/alter_from_regional_by_table_to_rbr
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr712218057
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_drop_region564512728
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterSurvivalGoals/alter_populated_database_from_zone_to_region-16 1 359332916 ns/op 41.00 roundtrips
BenchmarkAlterSurvivalGoals/alter_populated_database_from_region_to_zone
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone141752125
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region085553020
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr356440840
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone849513004
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr215803079
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
rtt_analysis_bench.go:131: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region053354155
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone938059163
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr701030222
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region037082402
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone648673362
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterRegions/alter_populated_database_drop_region-16 1 15275801616 ns/op 21.00 roundtrips
BenchmarkAlterRegions/alter_populated_database_add_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_add_region041894743
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region-16 1 325849061 ns/op 21.00 roundtrips
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=BenchmarkAlterPrimaryRegion PKG=./pkg/ccl/benchccl/rttanalysisccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience pawalt
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*BenchmarkAlterPrimaryRegion.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_priority | ccl benchccl rttanalysisccl benchmarkalterprimaryregion failed ccl benchccl rttanalysisccl benchmarkalterprimaryregion with on master benchmarkalterprimaryregion alter populated database alter primary region test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltertablelocality alter from global to rbr ns op roundtrips benchmarkaltertablelocality alter from regional by table to rbr test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterregions alter populated database drop test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltersurvivalgoals alter populated database from zone to region ns op roundtrips benchmarkaltersurvivalgoals alter populated database from region to zone test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterregions alter populated database drop region ns op roundtrips benchmarkalterregions alter populated database add region test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterregions alter populated database add test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterprimaryregion alter populated database alter primary region ns op roundtrips reproduce to reproduce try bash make stressrace tests benchmarkalterprimaryregion pkg pkg ccl benchccl rttanalysisccl testtimeout stressflags timeout parameters in this failure goflags json cc cockroachdb sql experience pawalt | 0 |
26,230 | 4,211,605,636 | IssuesEvent | 2016-06-29 14:02:34 | e-government-ua/i | https://api.github.com/repos/e-government-ua/i | closed | На дашборде реализовать "Интерфейс для работы с правилами эскалаций" | active question test _dashboard-js | ориентир:
https://test.region.igov.org.ua/escalations
логин/пароль: kermit/kermit
Только на основании своего перечня полей и сущностей, которые нужно туда "вывести":
- [x] 1) основная закладка со списком правил и возможностью их добавить, редактировать, удалять и клонировать (как в образце по расписаниям электронной очереди)
EscalationRule - правила
nID;sID_BP;sID_UserTask;sCondition;soData;sPatternFile;nID_EscalationRuleFunction
1;dnepr_dms_passport;"*";"nElapsedDays == 100";"{nDays:10, asRecipientMail:['bvv4ik@gmail.com', 'askosyr@gmail.com']}";"escalation/escalation_template.html";1
где:
sID_BP - ИД БизнесПроцесса (для удобства стоит отображать его название)
sID_UserTask;sCondition - формула условия
soData - JSON-обьект с параметрами (для удобства его лучше отображать в виде комбобокса, при віборе пунктов которого давать их править + редактировать)
sPatternFile - название файла с шаблоном основного сообщения
nID_EscalationRuleFunction - ИД функции(по сути комбобокс со списком из нижних сущностей, с названиями из sName)
- [x] 2) вспомогательная закладка со списком функций, с теми-же возможностями что и в п.1.
EscalationRuleFunction - функции (действия при выполнения правил)
nID;sName;sBeanHandler
1;Отправка уведомления на электронную почту;EscalationHandler_SendMailAlert
где:
sName - название
sBeanHandler - название Бина-обработчика
| 1.0 | На дашборде реализовать "Интерфейс для работы с правилами эскалаций" - ориентир:
https://test.region.igov.org.ua/escalations
логин/пароль: kermit/kermit
Только на основании своего перечня полей и сущностей, которые нужно туда "вывести":
- [x] 1) основная закладка со списком правил и возможностью их добавить, редактировать, удалять и клонировать (как в образце по расписаниям электронной очереди)
EscalationRule - правила
nID;sID_BP;sID_UserTask;sCondition;soData;sPatternFile;nID_EscalationRuleFunction
1;dnepr_dms_passport;"*";"nElapsedDays == 100";"{nDays:10, asRecipientMail:['bvv4ik@gmail.com', 'askosyr@gmail.com']}";"escalation/escalation_template.html";1
где:
sID_BP - ИД БизнесПроцесса (для удобства стоит отображать его название)
sID_UserTask;sCondition - формула условия
soData - JSON-обьект с параметрами (для удобства его лучше отображать в виде комбобокса, при віборе пунктов которого давать их править + редактировать)
sPatternFile - название файла с шаблоном основного сообщения
nID_EscalationRuleFunction - ИД функции(по сути комбобокс со списком из нижних сущностей, с названиями из sName)
- [x] 2) вспомогательная закладка со списком функций, с теми-же возможностями что и в п.1.
EscalationRuleFunction - функции (действия при выполнения правил)
nID;sName;sBeanHandler
1;Отправка уведомления на электронную почту;EscalationHandler_SendMailAlert
где:
sName - название
sBeanHandler - название Бина-обработчика
| non_priority | на дашборде реализовать интерфейс для работы с правилами эскалаций ориентир логин пароль kermit kermit только на основании своего перечня полей и сущностей которые нужно туда вывести основная закладка со списком правил и возможностью их добавить редактировать удалять и клонировать как в образце по расписаниям электронной очереди escalationrule правила nid sid bp sid usertask scondition sodata spatternfile nid escalationrulefunction dnepr dms passport nelapseddays ndays asrecipientmail escalation escalation template html где sid bp ид бизнеспроцесса для удобства стоит отображать его название sid usertask scondition формула условия sodata json обьект с параметрами для удобства его лучше отображать в виде комбобокса при віборе пунктов которого давать их править редактировать spatternfile название файла с шаблоном основного сообщения nid escalationrulefunction ид функции по сути комбобокс со списком из нижних сущностей с названиями из sname вспомогательная закладка со списком функций с теми же возможностями что и в п escalationrulefunction функции действия при выполнения правил nid sname sbeanhandler отправка уведомления на электронную почту escalationhandler sendmailalert где sname название sbeanhandler название бина обработчика | 0 |
13,240 | 15,707,747,013 | IssuesEvent | 2021-03-26 19:22:38 | correctcomputation/checkedc-clang | https://api.github.com/repos/correctcomputation/checkedc-clang | opened | `ClangTool::run` `chdir` call corrupts internal Clang include file path cache (?) | Upstream bug clang preprocessor command-line | As part of the change to expand macros before running 3C, I tried to change `convert_project` so that instead of (1) passing an adjusted version of the union of all compiler options seen in the compilation database to 3C via `-extra-arg-before`, it (2) lets 3C read the options directly from the compilation database. This is because approach (1) may be wrong if different translation units have different compiler options, and I was more concerned about this as `convert_project` started to have more direct interaction with the preprocessor. Importantly, the adjustment in (1) included expanding relative paths in `-I` options to absolute paths based on the working directory of the translation unit carrying the options.
Unfortunately, this change seemed to cause our icecast benchmark to trigger a bug in Clang LibTooling. The symptom looks like this:
```
2021-03-26 14:52:27.200 INFO generate_ccommands - run3C: Running:/home/matt/3c-3.wt/build/bin/3c -dump-stats -p /home/matt/benchmarks/icecast-2.4.4/compile_commands.json -extra-arg=-w -base-dir="/home/matt/benchmarks/icecast-2.4.4" -output-dir="/home/matt/benchmarks/icecast-2.4.4/out.checked" /home/matt/benchmarks/icecast-2.4.4/src/format_flac.c /home/matt/benchmarks/icecast-2.4.4/src/format_ogg.c /home/matt/benchmarks/icecast-2.4.4/src/format_kate.c /home/matt/benchmarks/icecast-2.4.4/src/main.c /home/matt/benchmarks/icecast-2.4.4/src/format_mp3.c /home/matt/benchmarks/icecast-2.4.4/src/sighandler.c /home/matt/benchmarks/icecast-2.4.4/src/global.c /home/matt/benchmarks/icecast-2.4.4/src/cfgfile.c /home/matt/benchmarks/icecast-2.4.4/src/format_ebml.c /home/matt/benchmarks/icecast-2.4.4/src/event.c /home/matt/benchmarks/icecast-2.4.4/src/auth_htpasswd.c /home/matt/benchmarks/icecast-2.4.4/src/refbuf.c /home/matt/benchmarks/icecast-2.4.4/src/avl/avl.c /home/matt/benchmarks/icecast-2.4.4/src/format_vorbis.c /home/matt/benchmarks/icecast-2.4.4/src/connection.c /home/matt/benchmarks/icecast-2.4.4/src/util.c /home/matt/benchmarks/icecast-2.4.4/src/admin.c /home/matt/benchmarks/icecast-2.4.4/src/log/log.c /home/matt/benchmarks/icecast-2.4.4/src/format_opus.c /home/matt/benchmarks/icecast-2.4.4/src/thread/thread.c /home/matt/benchmarks/icecast-2.4.4/src/client.c /home/matt/benchmarks/icecast-2.4.4/src/timing/timing.c /home/matt/benchmarks/icecast-2.4.4/src/net/resolver.c /home/matt/benchmarks/icecast-2.4.4/src/stats.c /home/matt/benchmarks/icecast-2.4.4/src/net/sock.c /home/matt/benchmarks/icecast-2.4.4/src/source.c /home/matt/benchmarks/icecast-2.4.4/src/slave.c /home/matt/benchmarks/icecast-2.4.4/src/format_skeleton.c /home/matt/benchmarks/icecast-2.4.4/src/logging.c /home/matt/benchmarks/icecast-2.4.4/src/fserve.c /home/matt/benchmarks/icecast-2.4.4/src/auth.c /home/matt/benchmarks/icecast-2.4.4/src/format_midi.c /home/matt/benchmarks/icecast-2.4.4/src/md5.c /home/matt/benchmarks/icecast-2.4.4/src/format.c /home/matt/benchmarks/icecast-2.4.4/src/xslt.c /home/matt/benchmarks/icecast-2.4.4/src/httpp/httpp.c
avl.c:33:11: fatal error: cannot open file '../config.h': No such file or directory
#include <config.h>
^
avl.c:33:11: fatal error: cannot open file '../config.h': No such file or directory
#include <config.h>
^
[...more similar errors...]
```
My rough theory is as follows: Clang has a cache where the first time it sees `#include STR` (where `STR` is of the form `<PATH>` or `"PATH"`), it searches the include path for the first matching file and caches the path at which it found the file (to a first approximation, the concatenation of the `-I` directory with `PATH`). If Clang later sees `#include STR` again, it tries to open the cached path directly and raises a fatal error (seen above) if it fails. The problem arises when the cached path is relative, which can occur if the directory path specified via `-I` was relative. `ClangTool::run` iterates over the specified translation units, and for each one, it does a `chdir` to the working directory specified in the compilation database but (apparently) does not invalidate the cache. Consequently, if different translation units have different working directories, the preprocessor may try and fail to open a cached relative path because the working directory is different than it was when the path was added to the cache, when instead the preprocessor should do the include search over. Surprisingly, #488 did not fix the problem because `ClangTool::buildASTs` still calls `ClangTool::run` internally (!).
Here is [the original benchmark workflow run in which the problem appeared](https://github.com/correctcomputation/actions/runs/2203501812?check_suite_focus=true) (though the logs will probably expire from GitHub soon). It should be possible to reproduce the problem by re-running that revision of the `preprocess-before-conversion` workflow (https://github.com/correctcomputation/actions/commit/7651529638585a2c23f576ee08398ede5af6f239) on the corresponding revision of the `preprocess-before-conversion` branch of this repository (c113b1dbf76bd51d0b5326594b47ee9f8e12a686). We could probably construct a smaller test case with a compilation database with two entries (and presumably that's what we would do if we wanted to add a regression test for the problem to 3C), but I don't want to take the time to do that now.
In a web search, I found a few reports of similar-looking problems ([1](https://lists.llvm.org/pipermail/cfe-dev/2013-December/033883.html), [2](https://lists.llvm.org/pipermail/cfe-users/2015-April/000681.html)), but it doesn't appear that anyone has tracked down the details and formally reported the bug in [the Clang bug database](https://bugs.llvm.org/). We could do so if we wish.
Ultimately, we'll probably want to fix or work around this problem somehow so that end users get correct behavior when running 3C on a compilation database like that of icecast. For now, I'm planning to work around the problem in `convert_project` by restoring the legacy behavior of passing `-extra-arg-before` to 3C, but only for the absolute versions of `-I` options. Since we use `-extra-arg-before`, this will ensure that every included file is found via an absolute `-I` directory before we reach the relative ones in the compilation database, so the cached path will be absolute, avoiding the problem. In principle, this could be wrong if different translation units have different sets of resolved `-I` directories: if we apply the union of the `-I` directories to all translation units, then a translation unit could use a file from an `-I` directory that was not supposed to be active for that translation unit, when it was intended to use a file from a later `-I` directory instead. However, I don't believe this happens in any of our current benchmarks. | 1.0 | `ClangTool::run` `chdir` call corrupts internal Clang include file path cache (?) - As part of the change to expand macros before running 3C, I tried to change `convert_project` so that instead of (1) passing an adjusted version of the union of all compiler options seen in the compilation database to 3C via `-extra-arg-before`, it (2) lets 3C read the options directly from the compilation database. This is because approach (1) may be wrong if different translation units have different compiler options, and I was more concerned about this as `convert_project` started to have more direct interaction with the preprocessor. Importantly, the adjustment in (1) included expanding relative paths in `-I` options to absolute paths based on the working directory of the translation unit carrying the options.
Unfortunately, this change seemed to cause our icecast benchmark to trigger a bug in Clang LibTooling. The symptom looks like this:
```
2021-03-26 14:52:27.200 INFO generate_ccommands - run3C: Running:/home/matt/3c-3.wt/build/bin/3c -dump-stats -p /home/matt/benchmarks/icecast-2.4.4/compile_commands.json -extra-arg=-w -base-dir="/home/matt/benchmarks/icecast-2.4.4" -output-dir="/home/matt/benchmarks/icecast-2.4.4/out.checked" /home/matt/benchmarks/icecast-2.4.4/src/format_flac.c /home/matt/benchmarks/icecast-2.4.4/src/format_ogg.c /home/matt/benchmarks/icecast-2.4.4/src/format_kate.c /home/matt/benchmarks/icecast-2.4.4/src/main.c /home/matt/benchmarks/icecast-2.4.4/src/format_mp3.c /home/matt/benchmarks/icecast-2.4.4/src/sighandler.c /home/matt/benchmarks/icecast-2.4.4/src/global.c /home/matt/benchmarks/icecast-2.4.4/src/cfgfile.c /home/matt/benchmarks/icecast-2.4.4/src/format_ebml.c /home/matt/benchmarks/icecast-2.4.4/src/event.c /home/matt/benchmarks/icecast-2.4.4/src/auth_htpasswd.c /home/matt/benchmarks/icecast-2.4.4/src/refbuf.c /home/matt/benchmarks/icecast-2.4.4/src/avl/avl.c /home/matt/benchmarks/icecast-2.4.4/src/format_vorbis.c /home/matt/benchmarks/icecast-2.4.4/src/connection.c /home/matt/benchmarks/icecast-2.4.4/src/util.c /home/matt/benchmarks/icecast-2.4.4/src/admin.c /home/matt/benchmarks/icecast-2.4.4/src/log/log.c /home/matt/benchmarks/icecast-2.4.4/src/format_opus.c /home/matt/benchmarks/icecast-2.4.4/src/thread/thread.c /home/matt/benchmarks/icecast-2.4.4/src/client.c /home/matt/benchmarks/icecast-2.4.4/src/timing/timing.c /home/matt/benchmarks/icecast-2.4.4/src/net/resolver.c /home/matt/benchmarks/icecast-2.4.4/src/stats.c /home/matt/benchmarks/icecast-2.4.4/src/net/sock.c /home/matt/benchmarks/icecast-2.4.4/src/source.c /home/matt/benchmarks/icecast-2.4.4/src/slave.c /home/matt/benchmarks/icecast-2.4.4/src/format_skeleton.c /home/matt/benchmarks/icecast-2.4.4/src/logging.c /home/matt/benchmarks/icecast-2.4.4/src/fserve.c /home/matt/benchmarks/icecast-2.4.4/src/auth.c /home/matt/benchmarks/icecast-2.4.4/src/format_midi.c /home/matt/benchmarks/icecast-2.4.4/src/md5.c /home/matt/benchmarks/icecast-2.4.4/src/format.c /home/matt/benchmarks/icecast-2.4.4/src/xslt.c /home/matt/benchmarks/icecast-2.4.4/src/httpp/httpp.c
avl.c:33:11: fatal error: cannot open file '../config.h': No such file or directory
#include <config.h>
^
avl.c:33:11: fatal error: cannot open file '../config.h': No such file or directory
#include <config.h>
^
[...more similar errors...]
```
My rough theory is as follows: Clang has a cache where the first time it sees `#include STR` (where `STR` is of the form `<PATH>` or `"PATH"`), it searches the include path for the first matching file and caches the path at which it found the file (to a first approximation, the concatenation of the `-I` directory with `PATH`). If Clang later sees `#include STR` again, it tries to open the cached path directly and raises a fatal error (seen above) if it fails. The problem arises when the cached path is relative, which can occur if the directory path specified via `-I` was relative. `ClangTool::run` iterates over the specified translation units, and for each one, it does a `chdir` to the working directory specified in the compilation database but (apparently) does not invalidate the cache. Consequently, if different translation units have different working directories, the preprocessor may try and fail to open a cached relative path because the working directory is different than it was when the path was added to the cache, when instead the preprocessor should do the include search over. Surprisingly, #488 did not fix the problem because `ClangTool::buildASTs` still calls `ClangTool::run` internally (!).
Here is [the original benchmark workflow run in which the problem appeared](https://github.com/correctcomputation/actions/runs/2203501812?check_suite_focus=true) (though the logs will probably expire from GitHub soon). It should be possible to reproduce the problem by re-running that revision of the `preprocess-before-conversion` workflow (https://github.com/correctcomputation/actions/commit/7651529638585a2c23f576ee08398ede5af6f239) on the corresponding revision of the `preprocess-before-conversion` branch of this repository (c113b1dbf76bd51d0b5326594b47ee9f8e12a686). We could probably construct a smaller test case with a compilation database with two entries (and presumably that's what we would do if we wanted to add a regression test for the problem to 3C), but I don't want to take the time to do that now.
In a web search, I found a few reports of similar-looking problems ([1](https://lists.llvm.org/pipermail/cfe-dev/2013-December/033883.html), [2](https://lists.llvm.org/pipermail/cfe-users/2015-April/000681.html)), but it doesn't appear that anyone has tracked down the details and formally reported the bug in [the Clang bug database](https://bugs.llvm.org/). We could do so if we wish.
Ultimately, we'll probably want to fix or work around this problem somehow so that end users get correct behavior when running 3C on a compilation database like that of icecast. For now, I'm planning to work around the problem in `convert_project` by restoring the legacy behavior of passing `-extra-arg-before` to 3C, but only for the absolute versions of `-I` options. Since we use `-extra-arg-before`, this will ensure that every included file is found via an absolute `-I` directory before we reach the relative ones in the compilation database, so the cached path will be absolute, avoiding the problem. In principle, this could be wrong if different translation units have different sets of resolved `-I` directories: if we apply the union of the `-I` directories to all translation units, then a translation unit could use a file from an `-I` directory that was not supposed to be active for that translation unit, when it was intended to use a file from a later `-I` directory instead. However, I don't believe this happens in any of our current benchmarks. | non_priority | clangtool run chdir call corrupts internal clang include file path cache as part of the change to expand macros before running i tried to change convert project so that instead of passing an adjusted version of the union of all compiler options seen in the compilation database to via extra arg before it lets read the options directly from the compilation database this is because approach may be wrong if different translation units have different compiler options and i was more concerned about this as convert project started to have more direct interaction with the preprocessor importantly the adjustment in included expanding relative paths in i options to absolute paths based on the working directory of the translation unit carrying the options unfortunately this change seemed to cause our icecast benchmark to trigger a bug in clang libtooling the symptom looks like this info generate ccommands running home matt wt build bin dump stats p home matt benchmarks icecast compile commands json extra arg w base dir home matt benchmarks icecast output dir home matt benchmarks icecast out checked home matt benchmarks icecast src format flac c home matt benchmarks icecast src format ogg c home matt benchmarks icecast src format kate c home matt benchmarks icecast src main c home matt benchmarks icecast src format c home matt benchmarks icecast src sighandler c home matt benchmarks icecast src global c home matt benchmarks icecast src cfgfile c home matt benchmarks icecast src format ebml c home matt benchmarks icecast src event c home matt benchmarks icecast src auth htpasswd c home matt benchmarks icecast src refbuf c home matt benchmarks icecast src avl avl c home matt benchmarks icecast src format vorbis c home matt benchmarks icecast src connection c home matt benchmarks icecast src util c home matt benchmarks icecast src admin c home matt benchmarks icecast src log log c home matt benchmarks icecast src format opus c home matt benchmarks icecast src thread thread c home matt benchmarks icecast src client c home matt benchmarks icecast src timing timing c home matt benchmarks icecast src net resolver c home matt benchmarks icecast src stats c home matt benchmarks icecast src net sock c home matt benchmarks icecast src source c home matt benchmarks icecast src slave c home matt benchmarks icecast src format skeleton c home matt benchmarks icecast src logging c home matt benchmarks icecast src fserve c home matt benchmarks icecast src auth c home matt benchmarks icecast src format midi c home matt benchmarks icecast src c home matt benchmarks icecast src format c home matt benchmarks icecast src xslt c home matt benchmarks icecast src httpp httpp c avl c fatal error cannot open file config h no such file or directory include avl c fatal error cannot open file config h no such file or directory include my rough theory is as follows clang has a cache where the first time it sees include str where str is of the form or path it searches the include path for the first matching file and caches the path at which it found the file to a first approximation the concatenation of the i directory with path if clang later sees include str again it tries to open the cached path directly and raises a fatal error seen above if it fails the problem arises when the cached path is relative which can occur if the directory path specified via i was relative clangtool run iterates over the specified translation units and for each one it does a chdir to the working directory specified in the compilation database but apparently does not invalidate the cache consequently if different translation units have different working directories the preprocessor may try and fail to open a cached relative path because the working directory is different than it was when the path was added to the cache when instead the preprocessor should do the include search over surprisingly did not fix the problem because clangtool buildasts still calls clangtool run internally here is though the logs will probably expire from github soon it should be possible to reproduce the problem by re running that revision of the preprocess before conversion workflow on the corresponding revision of the preprocess before conversion branch of this repository we could probably construct a smaller test case with a compilation database with two entries and presumably that s what we would do if we wanted to add a regression test for the problem to but i don t want to take the time to do that now in a web search i found a few reports of similar looking problems but it doesn t appear that anyone has tracked down the details and formally reported the bug in we could do so if we wish ultimately we ll probably want to fix or work around this problem somehow so that end users get correct behavior when running on a compilation database like that of icecast for now i m planning to work around the problem in convert project by restoring the legacy behavior of passing extra arg before to but only for the absolute versions of i options since we use extra arg before this will ensure that every included file is found via an absolute i directory before we reach the relative ones in the compilation database so the cached path will be absolute avoiding the problem in principle this could be wrong if different translation units have different sets of resolved i directories if we apply the union of the i directories to all translation units then a translation unit could use a file from an i directory that was not supposed to be active for that translation unit when it was intended to use a file from a later i directory instead however i don t believe this happens in any of our current benchmarks | 0 |
187,724 | 22,045,869,736 | IssuesEvent | 2022-05-30 01:35:01 | Nivaskumark/kernel_v4.1.15 | https://api.github.com/repos/Nivaskumark/kernel_v4.1.15 | closed | CVE-2018-16862 (Medium) detected in linuxlinux-4.6 - autoclosed | security vulnerability | ## CVE-2018-16862 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.1.15/commit/00db4e8795bcbec692fb60b19160bdd763ad42e3">00db4e8795bcbec692fb60b19160bdd763ad42e3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/truncate.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/truncate.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security flaw was found in the Linux kernel in a way that the cleancache subsystem clears an inode after the final file truncation (removal). The new file created with the same inode may contain leftover pages from cleancache and the old file data instead of the new one.
<p>Publish Date: 2018-11-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16862>CVE-2018-16862</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-16862">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-16862</a></p>
<p>Release Date: 2018-11-26</p>
<p>Fix Resolution: v4.20-rc5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-16862 (Medium) detected in linuxlinux-4.6 - autoclosed - ## CVE-2018-16862 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.1.15/commit/00db4e8795bcbec692fb60b19160bdd763ad42e3">00db4e8795bcbec692fb60b19160bdd763ad42e3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/truncate.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/truncate.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security flaw was found in the Linux kernel in a way that the cleancache subsystem clears an inode after the final file truncation (removal). The new file created with the same inode may contain leftover pages from cleancache and the old file data instead of the new one.
<p>Publish Date: 2018-11-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16862>CVE-2018-16862</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-16862">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-16862</a></p>
<p>Release Date: 2018-11-26</p>
<p>Fix Resolution: v4.20-rc5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files mm truncate c mm truncate c vulnerability details a security flaw was found in the linux kernel in a way that the cleancache subsystem clears an inode after the final file truncation removal the new file created with the same inode may contain leftover pages from cleancache and the old file data instead of the new one publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
8,525 | 11,965,103,792 | IssuesEvent | 2020-04-05 22:02:51 | DataScienceResearchPeru/covid-19_latinoamerica | https://api.github.com/repos/DataScienceResearchPeru/covid-19_latinoamerica | closed | Maps Covid-19 Latin America | Contributors Requirement Urgent | **Is your feature request related to a problem? Please describe.**
We need a map.
| 1.0 | Maps Covid-19 Latin America - **Is your feature request related to a problem? Please describe.**
We need a map.
| non_priority | maps covid latin america is your feature request related to a problem please describe we need a map | 0 |
22,072 | 6,228,304,556 | IssuesEvent | 2017-07-10 22:59:03 | XceedBoucherS/TestImport5 | https://api.github.com/repos/XceedBoucherS/TestImport5 | closed | Duplicate ResourceDictionary | CodePlex | <b>gavaiken[CodePlex]</b> <br />http://wpftoolkit.codeplex.com/SourceControl/changeset/view/83405#2006971
nbsp
I was wondering if this was intentional and if so, why? Sorry for not submitting a patch, I don't have the infrastructure currently.
| 1.0 | Duplicate ResourceDictionary - <b>gavaiken[CodePlex]</b> <br />http://wpftoolkit.codeplex.com/SourceControl/changeset/view/83405#2006971
nbsp
I was wondering if this was intentional and if so, why? Sorry for not submitting a patch, I don't have the infrastructure currently.
| non_priority | duplicate resourcedictionary gavaiken nbsp i was wondering if this was intentional and if so why sorry for not submitting a patch i don t have the infrastructure currently | 0 |
47,096 | 19,559,832,469 | IssuesEvent | 2022-01-03 14:49:02 | PreMiD/Presences | https://api.github.com/repos/PreMiD/Presences | opened | Strip Poker Night at the Inventory (SPNATI) | spnati.net | 🔞 NSFW Service Request | ### Discussed in https://github.com/PreMiD/Presences/discussions/4447
<div type='discussions-op-text'>
<sup>Originally posted by **fr0r** February 7, 2020</sup>
**Prerequisites and essential questions** <!--- Required, please answer the following questions as honestly as possible by changing the "[ ]" to "[x]" or by marking it after creating the issue (easier), not checking a question counts as "No". -->
* [x] Is it a popular site?
* [x] Is the website older than 2 months? <!--- It is necessary for the website to be older than 2 months. -->
* [ ] Is the site locked to a specific country/region?
* [ ] Is the site a paid service? (e.g. Netflix, Hulu)
* [x] Does the website feature NSFW content? (e.g. porn, etc...)
* [ ] Are you a donator/patron?
* [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority?
**What's your Discord username?** <!--- Optional, unless you are a donator/patron. Ex. Clyde#0000 -->
fror#9493
**What's the name of the service?** <!--- Required, Ex. www.youtube.com | YouTube -->
Strip Poker Night at the Inventory | spnati.net
**What should the Presence display?** <!--- Required, make sure to be as clear as possible on what should be added. -->
Information about the round and status of my character
**If possible, please provide a logo for the service (512x512 minimum)** <!--- Optional, it is recommended to upload the image here instead of using a 3rd-party host. -->
</div> | 1.0 | Strip Poker Night at the Inventory (SPNATI) | spnati.net - ### Discussed in https://github.com/PreMiD/Presences/discussions/4447
<div type='discussions-op-text'>
<sup>Originally posted by **fr0r** February 7, 2020</sup>
**Prerequisites and essential questions** <!--- Required, please answer the following questions as honestly as possible by changing the "[ ]" to "[x]" or by marking it after creating the issue (easier), not checking a question counts as "No". -->
* [x] Is it a popular site?
* [x] Is the website older than 2 months? <!--- It is necessary for the website to be older than 2 months. -->
* [ ] Is the site locked to a specific country/region?
* [ ] Is the site a paid service? (e.g. Netflix, Hulu)
* [x] Does the website feature NSFW content? (e.g. porn, etc...)
* [ ] Are you a donator/patron?
* [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority?
**What's your Discord username?** <!--- Optional, unless you are a donator/patron. Ex. Clyde#0000 -->
fror#9493
**What's the name of the service?** <!--- Required, Ex. www.youtube.com | YouTube -->
Strip Poker Night at the Inventory | spnati.net
**What should the Presence display?** <!--- Required, make sure to be as clear as possible on what should be added. -->
Information about the round and status of my character
**If possible, please provide a logo for the service (512x512 minimum)** <!--- Optional, it is recommended to upload the image here instead of using a 3rd-party host. -->
</div> | non_priority | strip poker night at the inventory spnati spnati net discussed in originally posted by february prerequisites and essential questions is it a popular site is the website older than months is the site locked to a specific country region is the site a paid service e g netflix hulu does the website feature nsfw content e g porn etc are you a donator patron do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority what s your discord username fror what s the name of the service strip poker night at the inventory spnati net what should the presence display information about the round and status of my character if possible please provide a logo for the service minimum | 0 |
8,360 | 22,149,317,184 | IssuesEvent | 2022-06-03 15:09:16 | ActiveLogin/ActiveLogin.Authentication | https://api.github.com/repos/ActiveLogin/ActiveLogin.Authentication | closed | Implement UI for BankID Signing | enhancement BankID NewArchitecture | **Is your feature request related to a problem? Please describe.**
BankID supports Sign in addition to Login. At the moment we do have support in the API wrapper, but no UI support. By providing UI support users could sign things more easy.
**What area is it related to**
BankID
**Describe the solution you'd like**
As signing is not as standarized as login, I'm not sure. Let's use this issue to discuss possible implementations. | 1.0 | Implement UI for BankID Signing - **Is your feature request related to a problem? Please describe.**
BankID supports Sign in addition to Login. At the moment we do have support in the API wrapper, but no UI support. By providing UI support users could sign things more easy.
**What area is it related to**
BankID
**Describe the solution you'd like**
As signing is not as standarized as login, I'm not sure. Let's use this issue to discuss possible implementations. | non_priority | implement ui for bankid signing is your feature request related to a problem please describe bankid supports sign in addition to login at the moment we do have support in the api wrapper but no ui support by providing ui support users could sign things more easy what area is it related to bankid describe the solution you d like as signing is not as standarized as login i m not sure let s use this issue to discuss possible implementations | 0 |
78,789 | 10,085,170,963 | IssuesEvent | 2019-07-25 17:26:39 | safuya/books | https://api.github.com/repos/safuya/books | opened | Add a code of conduct | bug documentation | If this becomes a project that others contribute to, I would like a code of conduct to make sure everyone treats each other well. | 1.0 | Add a code of conduct - If this becomes a project that others contribute to, I would like a code of conduct to make sure everyone treats each other well. | non_priority | add a code of conduct if this becomes a project that others contribute to i would like a code of conduct to make sure everyone treats each other well | 0 |
281,335 | 24,384,761,571 | IssuesEvent | 2022-10-04 10:47:14 | pterm/pterm | https://api.github.com/repos/pterm/pterm | closed | Disable macOS CI tests | proposal good first issue CI Hacktoberfest tests proposal-accepted | macOS should no longer be tested in CI. We have so many tests that capture the console output, that macOS runs out of file descriptors and fails. This is done by removing macOS from the matrix set in the workflow (GitHub Actions) file. | 1.0 | Disable macOS CI tests - macOS should no longer be tested in CI. We have so many tests that capture the console output, that macOS runs out of file descriptors and fails. This is done by removing macOS from the matrix set in the workflow (GitHub Actions) file. | non_priority | disable macos ci tests macos should no longer be tested in ci we have so many tests that capture the console output that macos runs out of file descriptors and fails this is done by removing macos from the matrix set in the workflow github actions file | 0 |
27,541 | 13,285,861,276 | IssuesEvent | 2020-08-24 08:49:34 | Altinn/altinn-studio | https://api.github.com/repos/Altinn/altinn-studio | opened | Re-order request during instantiation | kind/user-story ops/performance solution/app-frontend | ## Description
During instantiation some requests from frontend are triggered after instantiation is complete.
By sending these request during instantiation we can reduce the instatiation time.
Could /api/metadata/ServiceMetaData and /api/jsonschema/default be sent during the instantiation request?
Once it is confirmed that party may instantiate, we should be fine to retrieve this data.
## Screenshots
> Screenshots or links to Figma (make sure your sketch is public)

## Acceptance criteria
The two requests are triggered before or during POST instance.
## Development tasks
- [ ] Change order of requests in app fronten.
- [ ] QA
- [ ] Manual test is complete (if relevant)
| True | Re-order request during instantiation - ## Description
During instantiation some requests from frontend are triggered after instantiation is complete.
By sending these request during instantiation we can reduce the instatiation time.
Could /api/metadata/ServiceMetaData and /api/jsonschema/default be sent during the instantiation request?
Once it is confirmed that party may instantiate, we should be fine to retrieve this data.
## Screenshots
> Screenshots or links to Figma (make sure your sketch is public)

## Acceptance criteria
The two requests are triggered before or during POST instance.
## Development tasks
- [ ] Change order of requests in app fronten.
- [ ] QA
- [ ] Manual test is complete (if relevant)
| non_priority | re order request during instantiation description during instantiation some requests from frontend are triggered after instantiation is complete by sending these request during instantiation we can reduce the instatiation time could api metadata servicemetadata and api jsonschema default be sent during the instantiation request once it is confirmed that party may instantiate we should be fine to retrieve this data screenshots screenshots or links to figma make sure your sketch is public acceptance criteria the two requests are triggered before or during post instance development tasks change order of requests in app fronten qa manual test is complete if relevant | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.