Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
34,355 | 29,513,219,644 | IssuesEvent | 2023-06-04 07:06:53 | bllendev/kalibre | https://api.github.com/repos/bllendev/kalibre | opened | react infra - Integrate React App with Django | infrastructure | - Set up the necessary configuration so that the React app can be served from the Django project. This can involve configuring Django to serve the React build files and setting up proxying for API requests during development. | 1.0 | react infra - Integrate React App with Django - - Set up the necessary configuration so that the React app can be served from the Django project. This can involve configuring Django to serve the React build files and setting up proxying for API requests during development. | infrastructure | react infra integrate react app with django set up the necessary configuration so that the react app can be served from the django project this can involve configuring django to serve the react build files and setting up proxying for api requests during development | 1 |
198,056 | 6,969,227,225 | IssuesEvent | 2017-12-11 03:44:57 | gw2efficiency/issues | https://api.github.com/repos/gw2efficiency/issues | closed | Gifts (for legendary armor) only visible when in bank | Bug Priority B Stalled: Discussion Needed | Hi.
If i place gifts (bones/dust etc) on my character, it is not showing on item search (and thus it is not calculated for crafting as 'owned material'. If i place this items in bank, they are found and properly calculated.
Browsing specific character inventory (not searching) properly shows item in question in their bag slots though.
It might be related to [this issue](https://github.com/gw2efficiency/issues/issues/752)
Items which i tested and are affected:
- Gift of Dust
- Gift of Bones
- Gift of Scales
- Gift of Fangs
- Gift of Blood
- Gift of Totems
- Gift of Claws
- Gift of Venom
- Gift of Dedication
- Eldritch Scroll
- Legendary Insight
Haven't tested more, all of those are used for Legendary Armor which i happend to be crafting when i noticed it. | 1.0 | Gifts (for legendary armor) only visible when in bank - Hi.
If i place gifts (bones/dust etc) on my character, it is not showing on item search (and thus it is not calculated for crafting as 'owned material'. If i place this items in bank, they are found and properly calculated.
Browsing specific character inventory (not searching) properly shows item in question in their bag slots though.
It might be related to [this issue](https://github.com/gw2efficiency/issues/issues/752)
Items which i tested and are affected:
- Gift of Dust
- Gift of Bones
- Gift of Scales
- Gift of Fangs
- Gift of Blood
- Gift of Totems
- Gift of Claws
- Gift of Venom
- Gift of Dedication
- Eldritch Scroll
- Legendary Insight
Haven't tested more, all of those are used for Legendary Armor which i happend to be crafting when i noticed it. | non_infrastructure | gifts for legendary armor only visible when in bank hi if i place gifts bones dust etc on my character it is not showing on item search and thus it is not calculated for crafting as owned material if i place this items in bank they are found and properly calculated browsing specific character inventory not searching properly shows item in question in their bag slots though it might be related to items which i tested and are affected gift of dust gift of bones gift of scales gift of fangs gift of blood gift of totems gift of claws gift of venom gift of dedication eldritch scroll legendary insight haven t tested more all of those are used for legendary armor which i happend to be crafting when i noticed it | 0 |
225,456 | 24,840,466,850 | IssuesEvent | 2022-10-26 12:20:38 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | aspnet:5.0-alpine docker image has reference to zlib-1.2.12-r0 which has vulnerability CVE-2022-37434 | Security | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
We are using docker image [mcr.microsoft.com/dotnet/aspnet:5.0-alpine] and we found out that it has a zlib-1.2.12-r0 package with vulnerability CVE-2022-37434.
We have confirmed that by reviewing the packages in this docker images using the command:
`docker run --rm mcr.microsoft.com/dotnet/aspnet:5.0-alpine apk list`

Checking the history of the docker image:
`docker image history --no-trunc mcr.microsoft.com/dotnet/aspnet:5.0-alpine`

we can see that this image hasn't been rebuilt for 5 months, and therefore it wouldn't have picked the latest zlib version (1.2.13) that has fixed the vulnerability
### Expected Behavior
The aspnet:5.0-alpine docker image should be rebuilt to get the latest fixed zlib package. Ideally this should have been done automatically
### Steps To Reproduce
We run aspnet:5.0-alpine docker image on Azure App Service for Containers, and the Azure Cloud Defender raised a vulnerability alert that it has a CVE-2022-37434 vulnerability
### Exceptions (if any)
ASP.Net docker images should be updated with the latest vulnerability fixes especially when the fixed packages have been available for over 2 months
### .NET Version
_No response_
### Anything else?
_No response_ | True | aspnet:5.0-alpine docker image has reference to zlib-1.2.12-r0 which has vulnerability CVE-2022-37434 - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
We are using docker image [mcr.microsoft.com/dotnet/aspnet:5.0-alpine] and we found out that it has a zlib-1.2.12-r0 package with vulnerability CVE-2022-37434.
We have confirmed that by reviewing the packages in this docker images using the command:
`docker run --rm mcr.microsoft.com/dotnet/aspnet:5.0-alpine apk list`

Checking the history of the docker image:
`docker image history --no-trunc mcr.microsoft.com/dotnet/aspnet:5.0-alpine`

we can see that this image hasn't been rebuilt for 5 months, and therefore it wouldn't have picked the latest zlib version (1.2.13) that has fixed the vulnerability
### Expected Behavior
The aspnet:5.0-alpine docker image should be rebuilt to get the latest fixed zlib package. Ideally this should have been done automatically
### Steps To Reproduce
We run aspnet:5.0-alpine docker image on Azure App Service for Containers, and the Azure Cloud Defender raised a vulnerability alert that it has a CVE-2022-37434 vulnerability
### Exceptions (if any)
ASP.Net docker images should be updated with the latest vulnerability fixes especially when the fixed packages have been available for over 2 months
### .NET Version
_No response_
### Anything else?
_No response_ | non_infrastructure | aspnet alpine docker image has reference to zlib which has vulnerability cve is there an existing issue for this i have searched the existing issues describe the bug we are using docker image and we found out that it has a zlib package with vulnerability cve we have confirmed that by reviewing the packages in this docker images using the command docker run rm mcr microsoft com dotnet aspnet alpine apk list checking the history of the docker image docker image history no trunc mcr microsoft com dotnet aspnet alpine we can see that this image hasn t been rebuilt for months and therefore it wouldn t have picked the latest zlib version that has fixed the vulnerability expected behavior the aspnet alpine docker image should be rebuilt to get the latest fixed zlib package ideally this should have been done automatically steps to reproduce we run aspnet alpine docker image on azure app service for containers and the azure cloud defender raised a vulnerability alert that it has a cve vulnerability exceptions if any asp net docker images should be updated with the latest vulnerability fixes especially when the fixed packages have been available for over months net version no response anything else no response | 0 |
10,251 | 8,452,988,349 | IssuesEvent | 2018-10-20 10:53:33 | TeamBravo2018/cloned-rfid-card-detection | https://api.github.com/repos/TeamBravo2018/cloned-rfid-card-detection | opened | Setup Message Broker | backlog item infrastructure messaging test production | ### Description ###
Setup HiveMq. Make sure that the deployments on PCF can communicate with the broker.
| 1.0 | Setup Message Broker - ### Description ###
Setup HiveMq. Make sure that the deployments on PCF can communicate with the broker.
| infrastructure | setup message broker description setup hivemq make sure that the deployments on pcf can communicate with the broker | 1 |
354,069 | 10,562,585,292 | IssuesEvent | 2019-10-04 18:42:20 | grpc/grpc | https://api.github.com/repos/grpc/grpc | reopened | MacOS Basic Tests C/C++ and Node Time Out | kind/bug lang/c++ lang/node priority/P1 | Basic Tests C/C++ MacOS and Basic Tests Node MacOS flakes.
When they fail, I see a bunch of messages saying "file has vanished"
> file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_qps_scenario_changes.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_shellcheck.sh"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_submodules.sh"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_test_filtering.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_tracer_sanity.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_version.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/core_banned_functions.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/core_untyped_structs.sh"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/sanity_tests.yaml"
rsync warning: some files vanished before they could be transferred (code 24) at main.c(1677) [generator=3.1.3]
https://source.cloud.google.com/results/invocations/9b4b867c-54ba-436a-b188-4e281627fe0d/targets
https://source.cloud.google.com/results/invocations/adb1dbdc-5999-4ad0-a2a0-2a6a2a4fb518/targets | 1.0 | MacOS Basic Tests C/C++ and Node Time Out - Basic Tests C/C++ MacOS and Basic Tests Node MacOS flakes.
When they fail, I see a bunch of messages saying "file has vanished"
> file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_qps_scenario_changes.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_shellcheck.sh"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_submodules.sh"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_test_filtering.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_tracer_sanity.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/check_version.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/core_banned_functions.py"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/core_untyped_structs.sh"
file has vanished: "/tmpfs/src/github/grpc/workspace_c_macos_opt_native/tools/run_tests/sanity/sanity_tests.yaml"
rsync warning: some files vanished before they could be transferred (code 24) at main.c(1677) [generator=3.1.3]
https://source.cloud.google.com/results/invocations/9b4b867c-54ba-436a-b188-4e281627fe0d/targets
https://source.cloud.google.com/results/invocations/adb1dbdc-5999-4ad0-a2a0-2a6a2a4fb518/targets | non_infrastructure | macos basic tests c c and node time out basic tests c c macos and basic tests node macos flakes when they fail i see a bunch of messages saying file has vanished file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity check qps scenario changes py file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity check shellcheck sh file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity check submodules sh file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity check test filtering py file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity check tracer sanity py file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity check version py file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity core banned functions py file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity core untyped structs sh file has vanished tmpfs src github grpc workspace c macos opt native tools run tests sanity sanity tests yaml rsync warning some files vanished before they could be transferred code at main c | 0 |
201,502 | 15,802,437,385 | IssuesEvent | 2021-04-03 09:44:08 | nowknowing/ped | https://api.github.com/repos/nowknowing/ped | opened | UG: add_person and add_booking lacking critical details | severity.High type.DocumentationBug | The mention of "multi-step" without further details of what the multi-step consists of is not acceptable.
Perhaps give at least one example of the "steps" i.e. commands that follow.
<!--session: 1617429926787-c5fe719d-2acf-4ed4-8116-9865b2720c99--> | 1.0 | UG: add_person and add_booking lacking critical details - The mention of "multi-step" without further details of what the multi-step consists of is not acceptable.
Perhaps give at least one example of the "steps" i.e. commands that follow.
<!--session: 1617429926787-c5fe719d-2acf-4ed4-8116-9865b2720c99--> | non_infrastructure | ug add person and add booking lacking critical details the mention of multi step without further details of what the multi step consists of is not acceptable perhaps give at least one example of the steps i e commands that follow | 0 |
15,840 | 11,727,540,313 | IssuesEvent | 2020-03-10 16:05:45 | reapit/foundations | https://api.github.com/repos/reapit/foundations | closed | Create Properties broker | feature infrastructure platform-team | Create properties broker function that fronts the properties service and offers facility to embed the following entities
- Images
- Documents
- Offers
- Negotiator
- Offices
- Department
- Vendor
- Area
Project and CI/CD pipelines should also be initialised as part of this ticket. Broker should become the entry point for the properties service which should no longer be directly public
The properties service should be updated to include the `embed` parameter option on the GET endpoints with the supported options listed above | 1.0 | Create Properties broker - Create properties broker function that fronts the properties service and offers facility to embed the following entities
- Images
- Documents
- Offers
- Negotiator
- Offices
- Department
- Vendor
- Area
Project and CI/CD pipelines should also be initialised as part of this ticket. Broker should become the entry point for the properties service which should no longer be directly public
The properties service should be updated to include the `embed` parameter option on the GET endpoints with the supported options listed above | infrastructure | create properties broker create properties broker function that fronts the properties service and offers facility to embed the following entities images documents offers negotiator offices department vendor area project and ci cd pipelines should also be initialised as part of this ticket broker should become the entry point for the properties service which should no longer be directly public the properties service should be updated to include the embed parameter option on the get endpoints with the supported options listed above | 1 |
11,161 | 8,969,351,239 | IssuesEvent | 2019-01-29 10:34:38 | GFDRR/open-risk-data-dashboard | https://api.github.com/repos/GFDRR/open-risk-data-dashboard | opened | Profile unnecessary fileds FE / BE | backend frontend infrastructure | The fileds First Name, Last Name, Title and Institution were removed from FE side (Registration form, My Profile & Admin User Profile manage).
In this moment in FE side the User profile Object contains this fileds because BE side requires them.
We must to remove all references to this fields from FE and BE side.
| 1.0 | Profile unnecessary fileds FE / BE - The fileds First Name, Last Name, Title and Institution were removed from FE side (Registration form, My Profile & Admin User Profile manage).
In this moment in FE side the User profile Object contains this fileds because BE side requires them.
We must to remove all references to this fields from FE and BE side.
| infrastructure | profile unnecessary fileds fe be the fileds first name last name title and institution were removed from fe side registration form my profile admin user profile manage in this moment in fe side the user profile object contains this fileds because be side requires them we must to remove all references to this fields from fe and be side | 1 |
6,309 | 6,311,744,469 | IssuesEvent | 2017-07-23 22:14:50 | twosigma/beakerx | https://api.github.com/repos/twosigma/beakerx | opened | set versions automatically | Enhancement Infrastructure | npm gets the version from package.json
and python gets it from _version.py
would be nice if they got it from git.
i think versioneer works for python, what about npm? | 1.0 | set versions automatically - npm gets the version from package.json
and python gets it from _version.py
would be nice if they got it from git.
i think versioneer works for python, what about npm? | infrastructure | set versions automatically npm gets the version from package json and python gets it from version py would be nice if they got it from git i think versioneer works for python what about npm | 1 |
279,541 | 24,233,458,469 | IssuesEvent | 2022-09-26 20:29:17 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | [Desktop] Clicking on Brave icon with no windows open should open a new window | tests OS/macOS closed/stale OS/Desktop | Carried over from https://github.com/brave/browser-laptop/issues/167
Test would do the following:
- [ ] have no windows open
- [ ] click on Brave icon in dock
- [ ] new window should be opened | 1.0 | [Desktop] Clicking on Brave icon with no windows open should open a new window - Carried over from https://github.com/brave/browser-laptop/issues/167
Test would do the following:
- [ ] have no windows open
- [ ] click on Brave icon in dock
- [ ] new window should be opened | non_infrastructure | clicking on brave icon with no windows open should open a new window carried over from test would do the following have no windows open click on brave icon in dock new window should be opened | 0 |
193,831 | 6,888,420,896 | IssuesEvent | 2017-11-22 05:49:05 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | Remove simplejson dependency from zulip | area: dependencies priority: high | Zulip still uses `simplejson` (Rather than the built-into-Python-3 `json` module or the even faster `ujson` module) for one thing: writing the `page_params` object into a `<script>` tag in `zerver/views/home.py`.
It'd be really great to be able to finish removing `simplejson` as a dependency. However, it is not trivial to do so, since we're using `simplejson.encoder.JSONEncoderForHTML`, and that doesn't seem to exist in `zerver/views/`
This is a great issue for someone who knows Python well, since I think a solution is likely to involve implementing a version of `simplejson.encoder.JSONEncoderForHTML` on top of the standard library.
| 1.0 | Remove simplejson dependency from zulip - Zulip still uses `simplejson` (Rather than the built-into-Python-3 `json` module or the even faster `ujson` module) for one thing: writing the `page_params` object into a `<script>` tag in `zerver/views/home.py`.
It'd be really great to be able to finish removing `simplejson` as a dependency. However, it is not trivial to do so, since we're using `simplejson.encoder.JSONEncoderForHTML`, and that doesn't seem to exist in `zerver/views/`
This is a great issue for someone who knows Python well, since I think a solution is likely to involve implementing a version of `simplejson.encoder.JSONEncoderForHTML` on top of the standard library.
| non_infrastructure | remove simplejson dependency from zulip zulip still uses simplejson rather than the built into python json module or the even faster ujson module for one thing writing the page params object into a tag in zerver views home py it d be really great to be able to finish removing simplejson as a dependency however it is not trivial to do so since we re using simplejson encoder jsonencoderforhtml and that doesn t seem to exist in zerver views this is a great issue for someone who knows python well since i think a solution is likely to involve implementing a version of simplejson encoder jsonencoderforhtml on top of the standard library | 0 |
3,406 | 4,290,594,489 | IssuesEvent | 2016-07-18 10:23:50 | Microsoft/visualfsharp | https://api.github.com/repos/Microsoft/visualfsharp | closed | Add jenkins build by branch and os | infrastructure | Current jenkins ci builds `master` branch in `Debug` and `Release` configuration in `Windows`.
Like http://dotnet-ci.cloudapp.net/job/Microsoft_visualfsharp/job/release_windows_nt/
Others useful build configuration:
- OS: Windows/Macosx/Linux, because are needed by xplat (coreclr)
- branch: `master`, `coreclr`, `vs2015` are active branches built by appveyor
Proposed name in jenkins `branch_conf_osname` . Actual is `conf_osname`
ref [dotnet jenkins new new branch model](https://github.com/dotnet/cli/pull/1386#issue-133290544)
/cc @otawfik-ms @TyOverby | 1.0 | Add jenkins build by branch and os - Current jenkins ci builds `master` branch in `Debug` and `Release` configuration in `Windows`.
Like http://dotnet-ci.cloudapp.net/job/Microsoft_visualfsharp/job/release_windows_nt/
Others useful build configuration:
- OS: Windows/Macosx/Linux, because are needed by xplat (coreclr)
- branch: `master`, `coreclr`, `vs2015` are active branches built by appveyor
Proposed name in jenkins `branch_conf_osname` . Actual is `conf_osname`
ref [dotnet jenkins new new branch model](https://github.com/dotnet/cli/pull/1386#issue-133290544)
/cc @otawfik-ms @TyOverby | infrastructure | add jenkins build by branch and os current jenkins ci builds master branch in debug and release configuration in windows like others useful build configuration os windows macosx linux because are needed by xplat coreclr branch master coreclr are active branches built by appveyor proposed name in jenkins branch conf osname actual is conf osname ref cc otawfik ms tyoverby | 1 |
134,315 | 29,994,273,485 | IssuesEvent | 2023-06-26 03:13:31 | FerretDB/FerretDB | https://api.github.com/repos/FerretDB/FerretDB | opened | Support `maxPoolSize` and `minPoolSize` connection option | code/chore not ready | ### What should be done?
https://www.mongodb.com/docs/manual/reference/connection-string/#mongodb-urioption-urioption.maxPoolSize
https://www.mongodb.com/docs/manual/reference/connection-string/#mongodb-urioption-urioption.minPoolSize
See the issue number added in the PR https://github.com/FerretDB/FerretDB/pull/2878
### Where?
https://github.com/FerretDB/FerretDB/tree/main/internal/clientconn
### Definition of Done
- all handlers updated;
- unit tests added/updated;
- integration/compatibility tests added/updated;
- spot refactorings done;
- user documentation updated or an issue to create documentation created;
- something else?
| 1.0 | Support `maxPoolSize` and `minPoolSize` connection option - ### What should be done?
https://www.mongodb.com/docs/manual/reference/connection-string/#mongodb-urioption-urioption.maxPoolSize
https://www.mongodb.com/docs/manual/reference/connection-string/#mongodb-urioption-urioption.minPoolSize
See the issue number added in the PR https://github.com/FerretDB/FerretDB/pull/2878
### Where?
https://github.com/FerretDB/FerretDB/tree/main/internal/clientconn
### Definition of Done
- all handlers updated;
- unit tests added/updated;
- integration/compatibility tests added/updated;
- spot refactorings done;
- user documentation updated or an issue to create documentation created;
- something else?
| non_infrastructure | support maxpoolsize and minpoolsize connection option what should be done see the issue number added in the pr where definition of done all handlers updated unit tests added updated integration compatibility tests added updated spot refactorings done user documentation updated or an issue to create documentation created something else | 0 |
9,539 | 8,029,904,361 | IssuesEvent | 2018-07-27 17:40:19 | brave/browser-laptop | https://api.github.com/repos/brave/browser-laptop | closed | npm install fails with npm 5.4 and higher | dev-setup infrastructure stale | ### Description
`npm install` fails with:
```bash
Finished generating code
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\\test.exe
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\test.pdb (Full PDB)
+ cppunitlite@1.0.0
added 1 package in 4.797s
npm ERR! Cannot read property 'pause' of undefined
```
### Steps to Reproduce
1. update npm to 5.4
2. get code
3. run `npm install`
**Actual result:**
```bash
Finished generating code
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\\test.exe
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\test.pdb (Full PDB)
+ cppunitlite@1.0.0
added 1 package in 4.797s
npm ERR! Cannot read property 'pause' of undefined
```
**Expected result:**
success
**Reproduces how often:** [What percentage of the time does it reproduce?]
100%
### Brave Version
0.21.0
### Additional Information
temporarily had to downgrade npm to version 5.3.0
| 1.0 | npm install fails with npm 5.4 and higher - ### Description
`npm install` fails with:
```bash
Finished generating code
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\\test.exe
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\test.pdb (Full PDB)
+ cppunitlite@1.0.0
added 1 package in 4.797s
npm ERR! Cannot read property 'pause' of undefined
```
### Steps to Reproduce
1. update npm to 5.4
2. get code
3. run `npm install`
**Actual result:**
```bash
Finished generating code
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\\test.exe
test.vcxproj -> C:\dev\browser-laptop\node_modules\tracking-protection\node_modules\cppunitlite\build\Release\test.pdb (Full PDB)
+ cppunitlite@1.0.0
added 1 package in 4.797s
npm ERR! Cannot read property 'pause' of undefined
```
**Expected result:**
success
**Reproduces how often:** [What percentage of the time does it reproduce?]
100%
### Brave Version
0.21.0
### Additional Information
temporarily had to downgrade npm to version 5.3.0
| infrastructure | npm install fails with npm and higher description npm install fails with bash finished generating code test vcxproj c dev browser laptop node modules tracking protection node modules cppunitlite build release test exe test vcxproj c dev browser laptop node modules tracking protection node modules cppunitlite build release test pdb full pdb cppunitlite added package in npm err cannot read property pause of undefined steps to reproduce update npm to get code run npm install actual result bash finished generating code test vcxproj c dev browser laptop node modules tracking protection node modules cppunitlite build release test exe test vcxproj c dev browser laptop node modules tracking protection node modules cppunitlite build release test pdb full pdb cppunitlite added package in npm err cannot read property pause of undefined expected result success reproduces how often brave version additional information temporarily had to downgrade npm to version | 1 |
22,890 | 15,604,457,466 | IssuesEvent | 2021-03-19 03:58:08 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Tracking issue: enable more MSVC warnings | area-Infrastructure-coreclr in pr up-for-grabs | *Initial cost estimate*: 1 week
*Initial contacts*: @trylek, @GrabYourPitchforks, @janvorli
We enable the following warning codes (as error) on some internal runs as part of compliance checks. The warnings are reviewed by hand. The manual reviews always show that we're ok, but it would be nice if we could turn these on as part of the public build. There's a little bit of work involved here to make things compile cleanly. But it would save us the trouble of running through manual review every release.
```cmake
add_compile_options(/we4018)
add_compile_options(/we4055)
add_compile_options(/we4146)
add_compile_options(/we4242)
add_compile_options(/we4244)
add_compile_options(/we4267)
add_compile_options(/we4302)
add_compile_options(/we4308)
add_compile_options(/we4509)
add_compile_options(/we4510)
add_compile_options(/we4532)
add_compile_options(/we4533)
add_compile_options(/we4610)
add_compile_options(/we4611)
add_compile_options(/we4700)
add_compile_options(/we4701)
add_compile_options(/we4703)
add_compile_options(/we4789)
add_compile_options(/we4995)
add_compile_options(/we4996)
``` | 1.0 | Tracking issue: enable more MSVC warnings - *Initial cost estimate*: 1 week
*Initial contacts*: @trylek, @GrabYourPitchforks, @janvorli
We enable the following warning codes (as error) on some internal runs as part of compliance checks. The warnings are reviewed by hand. The manual reviews always show that we're ok, but it would be nice if we could turn these on as part of the public build. There's a little bit of work involved here to make things compile cleanly. But it would save us the trouble of running through manual review every release.
```cmake
add_compile_options(/we4018)
add_compile_options(/we4055)
add_compile_options(/we4146)
add_compile_options(/we4242)
add_compile_options(/we4244)
add_compile_options(/we4267)
add_compile_options(/we4302)
add_compile_options(/we4308)
add_compile_options(/we4509)
add_compile_options(/we4510)
add_compile_options(/we4532)
add_compile_options(/we4533)
add_compile_options(/we4610)
add_compile_options(/we4611)
add_compile_options(/we4700)
add_compile_options(/we4701)
add_compile_options(/we4703)
add_compile_options(/we4789)
add_compile_options(/we4995)
add_compile_options(/we4996)
``` | infrastructure | tracking issue enable more msvc warnings initial cost estimate week initial contacts trylek grabyourpitchforks janvorli we enable the following warning codes as error on some internal runs as part of compliance checks the warnings are reviewed by hand the manual reviews always show that we re ok but it would be nice if we could turn these on as part of the public build there s a little bit of work involved here to make things compile cleanly but it would save us the trouble of running through manual review every release cmake add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options add compile options | 1 |
23,254 | 16,011,757,327 | IssuesEvent | 2021-04-20 11:31:47 | TeamFranka/affinity | https://api.github.com/repos/TeamFranka/affinity | opened | Registration Email test | infrastructure reliability | We have a test to do a signup, let's improve it by:
- [ ] adding [mailhog based email backend to our docker-compose](https://kreuzwerker.de/post/e2e-testing-of-emails-in-mailhog-using-cypress)
- [ ] hook it up to the parse-platform in the docker-compose
- [ ] advance the tests to:
- [ ] check for the email-required toast-message
- [ ] open the email, find the link, click it
- [ ] refresh the app and make sure the toast is gone
- [ ] add a second tests
- [ ] check for the email-required toast-message
- [ ] use it to ask for the email again
- [ ] open that new email, find the link, click it
- [ ] refresh the app and make sure the toast is gone | 1.0 | Registration Email test - We have a test to do a signup, let's improve it by:
- [ ] adding [mailhog based email backend to our docker-compose](https://kreuzwerker.de/post/e2e-testing-of-emails-in-mailhog-using-cypress)
- [ ] hook it up to the parse-platform in the docker-compose
- [ ] advance the tests to:
- [ ] check for the email-required toast-message
- [ ] open the email, find the link, click it
- [ ] refresh the app and make sure the toast is gone
- [ ] add a second tests
- [ ] check for the email-required toast-message
- [ ] use it to ask for the email again
- [ ] open that new email, find the link, click it
- [ ] refresh the app and make sure the toast is gone | infrastructure | registration email test we have a test to do a signup let s improve it by adding hook it up to the parse platform in the docker compose advance the tests to check for the email required toast message open the email find the link click it refresh the app and make sure the toast is gone add a second tests check for the email required toast message use it to ask for the email again open that new email find the link click it refresh the app and make sure the toast is gone | 1 |
354,474 | 25,167,677,129 | IssuesEvent | 2022-11-10 22:35:15 | notaryproject/notation | https://api.github.com/repos/notaryproject/notation | closed | Plugin config directory varies based on OS and should be clearer with docs | documentation | the [os.UserConfigDir()](https://github.com/notaryproject/notation/blob/c61102cdb414a9aa55980bd113496abef3fcf784/pkg/config/path.go#L59) varies and therefore isn't the same based on OS.
i.e. on Ubuntu `~/.config/notation/plugins/azure-kv` works fine.
on a mac it is `~/Library/Application Support/notation/plugins` | 1.0 | Plugin config directory varies based on OS and should be clearer with docs - the [os.UserConfigDir()](https://github.com/notaryproject/notation/blob/c61102cdb414a9aa55980bd113496abef3fcf784/pkg/config/path.go#L59) varies and therefore isn't the same based on OS.
i.e. on Ubuntu `~/.config/notation/plugins/azure-kv` works fine.
on a mac it is `~/Library/Application Support/notation/plugins` | non_infrastructure | plugin config directory varies based on os and should be clearer with docs the varies and therefore isn t the same based on os i e on ubuntu config notation plugins azure kv works fine on a mac it is library application support notation plugins | 0 |
22,486 | 15,217,773,177 | IssuesEvent | 2021-02-17 17:00:03 | airyhq/airy | https://api.github.com/repos/airyhq/airy | opened | Re-evaluate the hostnames for Airy Core | cli infrastructure needs discussion | It will be more convenient for local deployment to have one hostname and instead of additional subdomains, have a path in the request.
This will impact the ingress and probably the path will only be needed for the `frontend` services. An example mapping of the requests:
```
api.airy.core/* -> airy.core/*
webhooks.airy.core/* -> airy.core/*
ui.airy.core/* -> airy.core/ui/*
chatplugin.airy.core/* -> airy.core/ui/chatplugin/*
tools.airy.core/* -> airy.core/tools/*
``` | 1.0 | Re-evaluate the hostnames for Airy Core - It will be more convenient for local deployment to have one hostname and instead of additional subdomains, have a path in the request.
This will impact the ingress and probably the path will only be needed for the `frontend` services. An example mapping of the requests:
```
api.airy.core/* -> airy.core/*
webhooks.airy.core/* -> airy.core/*
ui.airy.core/* -> airy.core/ui/*
chatplugin.airy.core/* -> airy.core/ui/chatplugin/*
tools.airy.core/* -> airy.core/tools/*
``` | infrastructure | re evaluate the hostnames for airy core it will be more convenient for local deployment to have one hostname and instead of additional subdomains have a path in the request this will impact the ingress and probably the path will only be needed for the frontend services an example mapping of the requests api airy core airy core webhooks airy core airy core ui airy core airy core ui chatplugin airy core airy core ui chatplugin tools airy core airy core tools | 1 |
18,964 | 13,181,067,673 | IssuesEvent | 2020-08-12 13:47:22 | pymor/pymor | https://api.github.com/repos/pymor/pymor | closed | Gitlab CI replaces branch refs for docs upload | bug infrastructure | turns 2020.1.1 into 2020-1-1 so I manually had to as et a symlink in the docs repo for the release just now | 1.0 | Gitlab CI replaces branch refs for docs upload - turns 2020.1.1 into 2020-1-1 so I manually had to as et a symlink in the docs repo for the release just now | infrastructure | gitlab ci replaces branch refs for docs upload turns into so i manually had to as et a symlink in the docs repo for the release just now | 1 |
29,430 | 24,007,621,770 | IssuesEvent | 2022-09-14 15:56:47 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | Infrastructure Monitoring: Migrate DataDog to GovCloud instance | Epic operations monitoring platform-initiative infrastructure crew-platform team-platform-infrastructure FY22-Q3 | **Note: This initiative should move to recently complete. @jhouse-solvd 9/14**
## Description
DataDog is moving to a dedicated installation in VA's GovCloud. The DOTS team has offered to help migrate dashboards, alerts, and monitors, but cannot migrate historical data.
## Details
- The DOTS team is migrating DataDog from its current SaaS instance (datadoghq.com) to GovCloud (ddog-gov.com)
- Authentication methods for this service will change (will be using Ablevets Okta)
- Authentication to Okta will be with a username and password
- DataDog engineers are available in the VA OIT DevOps Slack workspace to discuss technical details
- **All historic data will be lost and unable to be migrated**
- DataDog agents are being re-deployed to point to the new instance
- There will be a new subdomain: vagov.ddog-gov.com
- Once the account is migrated, the existing account will remain active for a short time but will be sunset by June 30, 2022
- The Platform Infrastructure Team is meeting w/ the DOTS team to gather add'l details
## Add'l considerations
- Authentication method will apply to the whole DataDog account, not per user.
- DataDog in GovCloud will still be accessible from anywhere (even off of the VA network
## Definition of done
- Platform Crew users are able to access the new instance
- DataDog agents are sending metrics to the new instance
- DataDog monitors are sending notifications
- to Slack
- to PagerDuty
- DataDog resources have been migrated to the new instance
- Documentation exists that highlights the following:
- How to request access to the new DataDog instance
- How to access the new DataDog instance
- How to migrate resources to the new DataDog instance | 2.0 | Infrastructure Monitoring: Migrate DataDog to GovCloud instance - **Note: This initiative should move to recently complete. @jhouse-solvd 9/14**
## Description
DataDog is moving to a dedicated installation in VA's GovCloud. The DOTS team has offered to help migrate dashboards, alerts, and monitors, but cannot migrate historical data.
## Details
- The DOTS team is migrating DataDog from its current SaaS instance (datadoghq.com) to GovCloud (ddog-gov.com)
- Authentication methods for this service will change (will be using Ablevets Okta)
- Authentication to Okta will be with a username and password
- DataDog engineers are available in the VA OIT DevOps Slack workspace to discuss technical details
- **All historic data will be lost and unable to be migrated**
- DataDog agents are being re-deployed to point to the new instance
- There will be a new subdomain: vagov.ddog-gov.com
- Once the account is migrated, the existing account will remain active for a short time but will be sunset by June 30, 2022
- The Platform Infrastructure Team is meeting w/ the DOTS team to gather add'l details
## Add'l considerations
- Authentication method will apply to the whole DataDog account, not per user.
- DataDog in GovCloud will still be accessible from anywhere (even off of the VA network
## Definition of done
- Platform Crew users are able to access the new instance
- DataDog agents are sending metrics to the new instance
- DataDog monitors are sending notifications
- to Slack
- to PagerDuty
- DataDog resources have been migrated to the new instance
- Documentation exists that highlights the following:
- How to request access to the new DataDog instance
- How to access the new DataDog instance
- How to migrate resources to the new DataDog instance | infrastructure | infrastructure monitoring migrate datadog to govcloud instance note this initiative should move to recently complete jhouse solvd description datadog is moving to a dedicated installation in va s govcloud the dots team has offered to help migrate dashboards alerts and monitors but cannot migrate historical data details the dots team is migrating datadog from its current saas instance datadoghq com to govcloud ddog gov com authentication methods for this service will change will be using ablevets okta authentication to okta will be with a username and password datadog engineers are available in the va oit devops slack workspace to discuss technical details all historic data will be lost and unable to be migrated datadog agents are being re deployed to point to the new instance there will be a new subdomain vagov ddog gov com once the account is migrated the existing account will remain active for a short time but will be sunset by june the platform infrastructure team is meeting w the dots team to gather add l details add l considerations authentication method will apply to the whole datadog account not per user datadog in govcloud will still be accessible from anywhere even off of the va network definition of done platform crew users are able to access the new instance datadog agents are sending metrics to the new instance datadog monitors are sending notifications to slack to pagerduty datadog resources have been migrated to the new instance documentation exists that highlights the following how to request access to the new datadog instance how to access the new datadog instance how to migrate resources to the new datadog instance | 1 |
26,607 | 20,325,109,096 | IssuesEvent | 2022-02-18 04:33:50 | pixiebrix/pixiebrix-extension | https://api.github.com/repos/pixiebrix/pixiebrix-extension | closed | Additional error patterns to exclude from Rollbar reporting | infrastructure developer experience | These errors are plentiful and clogging up our Rollbar tubes. I think they're generally benign errors from the messenger?
If they were having business impact we'd see them as errors in brick/extension point execution?
* No frame with id 453 in tab 81.
* The frame was removed.
* Extension context invalidated.
The last one might be fixed already. I'm only seeing error telemetry for it in Rollbar for 1.5.3. It's in the list of [CONNECTION_ERROR_PATTERNS](http://github.com/pixiebrix/pixiebrix-extension/blob/afe1c8d8608bf90c692413d0157ff33f7db8e1da/src/errors.ts#L269-L269)
| 1.0 | Additional error patterns to exclude from Rollbar reporting - These errors are plentiful and clogging up our Rollbar tubes. I think they're generally benign errors from the messenger?
If they were having business impact we'd see them as errors in brick/extension point execution?
* No frame with id 453 in tab 81.
* The frame was removed.
* Extension context invalidated.
The last one might be fixed already. I'm only seeing error telemetry for it in Rollbar for 1.5.3. It's in the list of [CONNECTION_ERROR_PATTERNS](http://github.com/pixiebrix/pixiebrix-extension/blob/afe1c8d8608bf90c692413d0157ff33f7db8e1da/src/errors.ts#L269-L269)
| infrastructure | additional error patterns to exclude from rollbar reporting these errors are plentiful and clogging up our rollbar tubes i think they re generally benign errors from the messenger if they were having business impact we d see them as errors in brick extension point execution no frame with id in tab the frame was removed extension context invalidated the last one might be fixed already i m only seeing error telemetry for it in rollbar for it s in the list of | 1 |
390,167 | 11,525,776,335 | IssuesEvent | 2020-02-15 10:55:24 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Challenge questions API does not work with scope based access control | Affected/5.10.0-Beta2 Priority/High Type/Bug | Steps to reproduce
As provided in identity.xml challenge question API is secured with following scopes
```
<Resource context="(.*)/api/users/v1/(.*)/challenges(.*)" secured="true" http-method="all">
<Permissions>/permission/admin/manage/identity</Permissions>
<Scopes>internal_identity_mgt_view</Scopes>
<Scopes>internal_identity_mgt_update</Scopes>
<Scopes>internal_identity_mgt_create</Scopes>
<Scopes>internal_identity_mgt_delete</Scopes>
</Resource>
<Resource context="(.*)/api/users/v1/(.*)/challenge-answers(.*)" secured="true" http-method="all">
<Permissions>/permission/admin/manage/identity</Permissions>
<Scopes>internal_identity_mgt_view</Scopes>
<Scopes>internal_identity_mgt_update</Scopes>
<Scopes>internal_identity_mgt_create</Scopes>
<Scopes>internal_identity_mgt_delete</Scopes>
</Resource>
```
Retrieve an access token for the scope **internal_identity_mgt_view**
Use the access token in bearer header and try to view the available challenge questions for a user using https://localhost:9443/t/wso2.com/api/users/v1/8022efd8-01cf-4f29-80fd-9e20641ff6bc/challenges
It returns 403 forbidden error | 1.0 | Challenge questions API does not work with scope based access control - Steps to reproduce
As provided in identity.xml challenge question API is secured with following scopes
```
<Resource context="(.*)/api/users/v1/(.*)/challenges(.*)" secured="true" http-method="all">
<Permissions>/permission/admin/manage/identity</Permissions>
<Scopes>internal_identity_mgt_view</Scopes>
<Scopes>internal_identity_mgt_update</Scopes>
<Scopes>internal_identity_mgt_create</Scopes>
<Scopes>internal_identity_mgt_delete</Scopes>
</Resource>
<Resource context="(.*)/api/users/v1/(.*)/challenge-answers(.*)" secured="true" http-method="all">
<Permissions>/permission/admin/manage/identity</Permissions>
<Scopes>internal_identity_mgt_view</Scopes>
<Scopes>internal_identity_mgt_update</Scopes>
<Scopes>internal_identity_mgt_create</Scopes>
<Scopes>internal_identity_mgt_delete</Scopes>
</Resource>
```
Retrieve an access token for the scope **internal_identity_mgt_view**
Use the access token in bearer header and try to view the available challenge questions for a user using https://localhost:9443/t/wso2.com/api/users/v1/8022efd8-01cf-4f29-80fd-9e20641ff6bc/challenges
It returns 403 forbidden error | non_infrastructure | challenge questions api does not work with scope based access control steps to reproduce as provided in identity xml challenge question api is secured with following scopes permission admin manage identity internal identity mgt view internal identity mgt update internal identity mgt create internal identity mgt delete permission admin manage identity internal identity mgt view internal identity mgt update internal identity mgt create internal identity mgt delete retrieve an access token for the scope internal identity mgt view use the access token in bearer header and try to view the available challenge questions for a user using it returns forbidden error | 0 |
13,266 | 10,171,670,742 | IssuesEvent | 2019-08-08 08:56:05 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Webgl teximage tests timing out flakily on Firefox linux | P3 area-infrastructure type-bug web-libraries | All the webgl tests dealing with textures seem to be timing out flakily on Firefox linux. Marking all tests in fast/canvas/webgl as Timeout, Pass to handle this.
examples of tests that have been seen timing out:
co19/LayoutTests/fast/canvas/webgl/texImage2DImageDataTest_t01
co19/LayoutTests/fast/canvas/webgl/gl-teximage_t01
co19/LayoutTests/fast/canvas/webgl/tex-image-and-sub-image-2d-with-video-rgba4444_t01
co19/LayoutTests/fast/canvas/webgl/tex-image-and-sub-image-2d-with-image-rgb565_t01
| 1.0 | Webgl teximage tests timing out flakily on Firefox linux - All the webgl tests dealing with textures seem to be timing out flakily on Firefox linux. Marking all tests in fast/canvas/webgl as Timeout, Pass to handle this.
examples of tests that have been seen timing out:
co19/LayoutTests/fast/canvas/webgl/texImage2DImageDataTest_t01
co19/LayoutTests/fast/canvas/webgl/gl-teximage_t01
co19/LayoutTests/fast/canvas/webgl/tex-image-and-sub-image-2d-with-video-rgba4444_t01
co19/LayoutTests/fast/canvas/webgl/tex-image-and-sub-image-2d-with-image-rgb565_t01
| infrastructure | webgl teximage tests timing out flakily on firefox linux all the webgl tests dealing with textures seem to be timing out flakily on firefox linux marking all tests in fast canvas webgl as timeout pass to handle this examples of tests that have been seen timing out layouttests fast canvas webgl layouttests fast canvas webgl gl teximage layouttests fast canvas webgl tex image and sub image with video layouttests fast canvas webgl tex image and sub image with image | 1 |
19,177 | 13,199,480,535 | IssuesEvent | 2020-08-14 05:59:39 | aodn/aodn-portal | https://api.github.com/repos/aodn/aodn-portal | opened | Portal is not displaying links from GN3 correctly on step 3 | T1 - IMOS T2 - O&M - Continuous Improvement T3 Information Infrastructure | This issue tracker is only for AODN Portal issues.
### Steps to reproduce the issue
1. Add https://portal.aodn.org.au/search?uuid=e850651b-d65d-495b-8182-5dde35919616 to the Portal
2. Proceed to Info tab on step 2 look at link to acoustic data viewer
3. Proceed to Step 3 look at link to acoustic data viewer
### Expected behaviour
The link on step 3 should be displayed like Acoustic Data Viewer (preview and download raw data) as it is on step 2
### Actual behaviour
It is displayed as simply acoustic.aodn.org.au
| 1.0 | Portal is not displaying links from GN3 correctly on step 3 - This issue tracker is only for AODN Portal issues.
### Steps to reproduce the issue
1. Add https://portal.aodn.org.au/search?uuid=e850651b-d65d-495b-8182-5dde35919616 to the Portal
2. Proceed to Info tab on step 2 look at link to acoustic data viewer
3. Proceed to Step 3 look at link to acoustic data viewer
### Expected behaviour
The link on step 3 should be displayed like Acoustic Data Viewer (preview and download raw data) as it is on step 2
### Actual behaviour
It is displayed as simply acoustic.aodn.org.au
| infrastructure | portal is not displaying links from correctly on step this issue tracker is only for aodn portal issues steps to reproduce the issue add to the portal proceed to info tab on step look at link to acoustic data viewer proceed to step look at link to acoustic data viewer expected behaviour the link on step should be displayed like acoustic data viewer preview and download raw data as it is on step actual behaviour it is displayed as simply acoustic aodn org au | 1 |
24,278 | 17,079,464,759 | IssuesEvent | 2021-07-08 01:28:53 | zowe/docs-site | https://api.github.com/repos/zowe/docs-site | closed | Enhancement - Dark Theme option | area: site-infrastructure type: enhancement | Add a button to site for switching between Regular Theme (light) and Dark Theme. We could use a package like this one (which didn't seem to work with our site when importing it in module.exports in config.js by default, just tried quickly) https://www.npmjs.com/package/vuepress-theme-dark-new
OR borrow ideas from other style sheets to define our own dark theme.
Site would switch CSS styles it reads from based on the state of the button.
Very low priority, just a cool idea, especially for a project focused on developers who I find often prefer dark themes. | 1.0 | Enhancement - Dark Theme option - Add a button to site for switching between Regular Theme (light) and Dark Theme. We could use a package like this one (which didn't seem to work with our site when importing it in module.exports in config.js by default, just tried quickly) https://www.npmjs.com/package/vuepress-theme-dark-new
OR borrow ideas from other style sheets to define our own dark theme.
Site would switch CSS styles it reads from based on the state of the button.
Very low priority, just a cool idea, especially for a project focused on developers who I find often prefer dark themes. | infrastructure | enhancement dark theme option add a button to site for switching between regular theme light and dark theme we could use a package like this one which didn t seem to work with our site when importing it in module exports in config js by default just tried quickly or borrow ideas from other style sheets to define our own dark theme site would switch css styles it reads from based on the state of the button very low priority just a cool idea especially for a project focused on developers who i find often prefer dark themes | 1 |
29,905 | 24,381,438,241 | IssuesEvent | 2022-10-04 08:11:03 | thinktecture/relayserver | https://api.github.com/repos/thinktecture/relayserver | opened | Consolidate logging | enhancement cosmetic infrastructure | Make logging similar to what is done in .NET:
One `LoggerExtensions` class with all `[LoggerMessage]` definitions on it. EventId is counted from 1 per logging extension class.
Extensions class is scoped per project or even per single class when there are a lot of messages.
Examples: See HttpConnectionDispatcher.Log and HttpConnectionManager.Log in SignalR Core. | 1.0 | Consolidate logging - Make logging similar to what is done in .NET:
One `LoggerExtensions` class with all `[LoggerMessage]` definitions on it. EventId is counted from 1 per logging extension class.
Extensions class is scoped per project or even per single class when there are a lot of messages.
Examples: See HttpConnectionDispatcher.Log and HttpConnectionManager.Log in SignalR Core. | infrastructure | consolidate logging make logging similar to what is done in net one loggerextensions class with all definitions on it eventid is counted from per logging extension class extensions class is scoped per project or even per single class when there are a lot of messages examples see httpconnectiondispatcher log and httpconnectionmanager log in signalr core | 1 |
17,238 | 12,260,326,934 | IssuesEvent | 2020-05-06 18:06:24 | NeuronRobotics/nrjavaserial | https://api.github.com/repos/NeuronRobotics/nrjavaserial | closed | Get Gradle to build with Java 1.5 compatibility | infrastructure | #21 asked us to compile for Java 1.5, and we've maintained that between then and now (I think). However, Gradle doesn't run on Java 1.5 – it has a minimum of 1.6. Before we build a new JAR for the 3.12.0 release, we need to figure out the configuration required to get Gradle to target 1.5 from a 1.6 compiler, or to get Gradle to invoke a 1.5 `javac` when running under 1.6.
| 1.0 | Get Gradle to build with Java 1.5 compatibility - #21 asked us to compile for Java 1.5, and we've maintained that between then and now (I think). However, Gradle doesn't run on Java 1.5 – it has a minimum of 1.6. Before we build a new JAR for the 3.12.0 release, we need to figure out the configuration required to get Gradle to target 1.5 from a 1.6 compiler, or to get Gradle to invoke a 1.5 `javac` when running under 1.6.
| infrastructure | get gradle to build with java compatibility asked us to compile for java and we ve maintained that between then and now i think however gradle doesn t run on java – it has a minimum of before we build a new jar for the release we need to figure out the configuration required to get gradle to target from a compiler or to get gradle to invoke a javac when running under | 1 |
30,919 | 25,169,276,242 | IssuesEvent | 2022-11-11 00:42:32 | astropy/astroquery | https://api.github.com/repos/astropy/astroquery | closed | ENH: Add linkcheck to CI | testing infrastructure | As suggested by @bsipocz in https://github.com/astropy/astroquery/pull/1784#discussion_r463344856 .
Might need pinning like https://github.com/astropy/astropy/blob/d9ad8d486e6f792595f7d40bc9951389e5e57992/tox.ini#L82-L86 and then do https://github.com/astropy/astropy/blob/d9ad8d486e6f792595f7d40bc9951389e5e57992/tox.ini#L115 in the `docs` directory. Adapt those to however the CI is set up if you don't use `tox`. | 1.0 | ENH: Add linkcheck to CI - As suggested by @bsipocz in https://github.com/astropy/astroquery/pull/1784#discussion_r463344856 .
Might need pinning like https://github.com/astropy/astropy/blob/d9ad8d486e6f792595f7d40bc9951389e5e57992/tox.ini#L82-L86 and then do https://github.com/astropy/astropy/blob/d9ad8d486e6f792595f7d40bc9951389e5e57992/tox.ini#L115 in the `docs` directory. Adapt those to however the CI is set up if you don't use `tox`. | infrastructure | enh add linkcheck to ci as suggested by bsipocz in might need pinning like and then do in the docs directory adapt those to however the ci is set up if you don t use tox | 1 |
32,512 | 26,747,553,160 | IssuesEvent | 2023-01-30 17:01:08 | opendatahub-io/odh-dashboard | https://api.github.com/repos/opendatahub-io/odh-dashboard | closed | [Feature Request]: Remove Python 2 dependency | kind/enhancement infrastructure priority/normal dependencies | ### Feature description
In order to improve our building process we should find a way to not require installing python 2:
https://github.com/opendatahub-io/odh-dashboard/blob/aaebc79eb2b24d7024e4e97e9b95b75e70eb764c/Dockerfile#L27
Python 2 is needed because it is a dependency of [node-gyp](https://github.com/opendatahub-io/odh-dashboard/blob/main/frontend/package-lock.json#L11757-L11783) which is required by [node-sass](https://github.com/opendatahub-io/odh-dashboard/blob/main/frontend/package-lock.json#L11862-L11884).
[Node-sass is deprecated](https://sass-lang.com/blog/libsass-is-deprecated), finding an alternative library may not require Python 2.
| 1.0 | [Feature Request]: Remove Python 2 dependency - ### Feature description
In order to improve our building process we should find a way to not require installing python 2:
https://github.com/opendatahub-io/odh-dashboard/blob/aaebc79eb2b24d7024e4e97e9b95b75e70eb764c/Dockerfile#L27
Python 2 is needed because it is a dependency of [node-gyp](https://github.com/opendatahub-io/odh-dashboard/blob/main/frontend/package-lock.json#L11757-L11783) which is required by [node-sass](https://github.com/opendatahub-io/odh-dashboard/blob/main/frontend/package-lock.json#L11862-L11884).
[Node-sass is deprecated](https://sass-lang.com/blog/libsass-is-deprecated), finding an alternative library may not require Python 2.
| infrastructure | remove python dependency feature description in order to improve our building process we should find a way to not require installing python python is needed because it is a dependency of which is required by finding an alternative library may not require python | 1 |
18,776 | 13,099,116,587 | IssuesEvent | 2020-08-03 20:54:02 | patternfly/patternfly-org | https://api.github.com/repos/patternfly/patternfly-org | closed | Implement new context switcher styles | infrastructure | @redallen commented on [Wed Nov 06 2019](https://github.com/patternfly/gatsby-theme-patternfly-org/issues/102)
Possibly implement this design by @mceledonia. The Core and React versions of the site would just have the dropdown disabled.

Framework - 14px secondary text color RH text medium
React/HTML - 24px primary text color RH text bold
View repo - 14px link text color (#06c) RH text regular
| 1.0 | Implement new context switcher styles - @redallen commented on [Wed Nov 06 2019](https://github.com/patternfly/gatsby-theme-patternfly-org/issues/102)
Possibly implement this design by @mceledonia. The Core and React versions of the site would just have the dropdown disabled.

Framework - 14px secondary text color RH text medium
React/HTML - 24px primary text color RH text bold
View repo - 14px link text color (#06c) RH text regular
| infrastructure | implement new context switcher styles redallen commented on possibly implement this design by mceledonia the core and react versions of the site would just have the dropdown disabled framework secondary text color rh text medium react html primary text color rh text bold view repo link text color rh text regular | 1 |
134,796 | 10,932,208,128 | IssuesEvent | 2019-11-23 16:07:24 | Students-of-the-city-of-Kostroma/Ray-of-hope | https://api.github.com/repos/Students-of-the-city-of-Kostroma/Ray-of-hope | opened | Протестировать макет публикации нового поста для WEB | O5 PR4 Post Sprint 7 Testing WEB-frontend | Epic #97
Story #98
Протестировать макеты постов в WEB, созданные в ходе выполнения задачи #116, на соответствие требованиям.
Поднимать баги по мере нахождения.
| 1.0 | Протестировать макет публикации нового поста для WEB - Epic #97
Story #98
Протестировать макеты постов в WEB, созданные в ходе выполнения задачи #116, на соответствие требованиям.
Поднимать баги по мере нахождения.
| non_infrastructure | протестировать макет публикации нового поста для web epic story протестировать макеты постов в web созданные в ходе выполнения задачи на соответствие требованиям поднимать баги по мере нахождения | 0 |
18,223 | 12,837,640,677 | IssuesEvent | 2020-07-07 16:05:54 | blockframes/blockframes | https://api.github.com/repos/blockframes/blockframes | closed | Setup Firebase Local UI Emulator | Infrastructure | Setup the Firebase Local UI Emulator
&
Run the app against this.
| 1.0 | Setup Firebase Local UI Emulator - Setup the Firebase Local UI Emulator
&
Run the app against this.
| infrastructure | setup firebase local ui emulator setup the firebase local ui emulator run the app against this | 1 |
5,302 | 5,557,638,810 | IssuesEvent | 2017-03-24 12:41:32 | datasciencebr/serenata-de-amor | https://api.github.com/repos/datasciencebr/serenata-de-amor | closed | Plugin to use Neo4j within Jupyter notebooks | hackathon infrastructure medium | This is not an issue that will generate code for Serenata de Amor, but will create a while new package.
Using graph database will help in Serenata de Ampr us and I ran into @nicolewhite's repo that makes great dataviz using Neo4j inside Jupyter notebooks (check this [example](http://nicolewhite.github.io/neo4j-jupyter/hello-world.html) and the main [repo](https://github.com/nicolewhite/neo4j-jupyter)).
My idea (and via tweet Nicole's said she's fine with that) is to pack her code into a package instalable via `pip`. Adding the module to Serenata de Amor repo and we can use the dataviz in out notebook. That's the idea. | 1.0 | Plugin to use Neo4j within Jupyter notebooks - This is not an issue that will generate code for Serenata de Amor, but will create a while new package.
Using graph database will help in Serenata de Ampr us and I ran into @nicolewhite's repo that makes great dataviz using Neo4j inside Jupyter notebooks (check this [example](http://nicolewhite.github.io/neo4j-jupyter/hello-world.html) and the main [repo](https://github.com/nicolewhite/neo4j-jupyter)).
My idea (and via tweet Nicole's said she's fine with that) is to pack her code into a package instalable via `pip`. Adding the module to Serenata de Amor repo and we can use the dataviz in out notebook. That's the idea. | infrastructure | plugin to use within jupyter notebooks this is not an issue that will generate code for serenata de amor but will create a while new package using graph database will help in serenata de ampr us and i ran into nicolewhite s repo that makes great dataviz using inside jupyter notebooks check this and the main my idea and via tweet nicole s said she s fine with that is to pack her code into a package instalable via pip adding the module to serenata de amor repo and we can use the dataviz in out notebook that s the idea | 1 |
8,706 | 7,573,419,161 | IssuesEvent | 2018-04-23 17:43:08 | dart-lang/site-www | https://api.github.com/repos/dart-lang/site-www | opened | Document how to link to sites, search for non-macro links | Infrastructure | I think we need a tiny wiki doc (or section of a doc) about how to link to various sites. E.g. I just approved a PR that links directly to webdev.dartlang.org, but it would be nice to be able to point contributors to a clear list of the macros to use instead. It'd be similar to https://github.com/dart-lang/site-www/wiki/Referring-to-API-docs. Maybe it would incorporate it.
(And, obviously, we should search for that webdev.dartlang.org link and fix it. It wasn't worth holding up the PR for.)
Not sure whether this doc ultimately belongs in site-shared or here.
/cc @chalin | 1.0 | Document how to link to sites, search for non-macro links - I think we need a tiny wiki doc (or section of a doc) about how to link to various sites. E.g. I just approved a PR that links directly to webdev.dartlang.org, but it would be nice to be able to point contributors to a clear list of the macros to use instead. It'd be similar to https://github.com/dart-lang/site-www/wiki/Referring-to-API-docs. Maybe it would incorporate it.
(And, obviously, we should search for that webdev.dartlang.org link and fix it. It wasn't worth holding up the PR for.)
Not sure whether this doc ultimately belongs in site-shared or here.
/cc @chalin | infrastructure | document how to link to sites search for non macro links i think we need a tiny wiki doc or section of a doc about how to link to various sites e g i just approved a pr that links directly to webdev dartlang org but it would be nice to be able to point contributors to a clear list of the macros to use instead it d be similar to maybe it would incorporate it and obviously we should search for that webdev dartlang org link and fix it it wasn t worth holding up the pr for not sure whether this doc ultimately belongs in site shared or here cc chalin | 1 |
29,087 | 23,709,188,084 | IssuesEvent | 2022-08-30 06:09:25 | UnitTestBot/UTBotJava | https://api.github.com/repos/UnitTestBot/UTBotJava | closed | Night Statistics Monitoring | enhancement infrastructure | **Description**
We want to develop and improve our product and, of course, there are some changes and its combinations which, according to some statistics, can make UTBot worse.
The main idea is collecting statistics after made changes. But it takes too long to collect statistics on a huge project to do it after each push into master. Thus, we will do it every night when no one makes changes.
Also we need to visualize collected statistics to easily analyze them.
| 1.0 | Night Statistics Monitoring - **Description**
We want to develop and improve our product and, of course, there are some changes and its combinations which, according to some statistics, can make UTBot worse.
The main idea is collecting statistics after made changes. But it takes too long to collect statistics on a huge project to do it after each push into master. Thus, we will do it every night when no one makes changes.
Also we need to visualize collected statistics to easily analyze them.
| infrastructure | night statistics monitoring description we want to develop and improve our product and of course there are some changes and its combinations which according to some statistics can make utbot worse the main idea is collecting statistics after made changes but it takes too long to collect statistics on a huge project to do it after each push into master thus we will do it every night when no one makes changes also we need to visualize collected statistics to easily analyze them | 1 |
8,681 | 7,558,925,878 | IssuesEvent | 2018-04-20 00:53:17 | connormlewis/idb | https://api.github.com/repos/connormlewis/idb | closed | Refine acceptance testing | infrastructure | Need to add acceptances tests that test searching, sorting, and filtering. | 1.0 | Refine acceptance testing - Need to add acceptances tests that test searching, sorting, and filtering. | infrastructure | refine acceptance testing need to add acceptances tests that test searching sorting and filtering | 1 |
12,335 | 9,708,890,720 | IssuesEvent | 2019-05-28 08:50:55 | ethersphere/go-ethereum | https://api.github.com/repos/ethersphere/go-ethereum | closed | Kubernetes: Add grafana dashboards to visualise disk usage | infrastructure | Moved from issue: ethereum/swarm-cluster#77
@nonsense wrote:
>We need to add disk i/o utilisation graphs to both persistent volumes and local vm disks, so that we track disk i/o usage on Swarm.
>We are having a lot of issues with it, with Swarm using 100% of available IOPS on current setup. It would be difficult to improve this and add benchmarking tests unless we can monitor it on K8s. | 1.0 | Kubernetes: Add grafana dashboards to visualise disk usage - Moved from issue: ethereum/swarm-cluster#77
@nonsense wrote:
>We need to add disk i/o utilisation graphs to both persistent volumes and local vm disks, so that we track disk i/o usage on Swarm.
>We are having a lot of issues with it, with Swarm using 100% of available IOPS on current setup. It would be difficult to improve this and add benchmarking tests unless we can monitor it on K8s. | infrastructure | kubernetes add grafana dashboards to visualise disk usage moved from issue ethereum swarm cluster nonsense wrote we need to add disk i o utilisation graphs to both persistent volumes and local vm disks so that we track disk i o usage on swarm we are having a lot of issues with it with swarm using of available iops on current setup it would be difficult to improve this and add benchmarking tests unless we can monitor it on | 1 |
240,134 | 18,293,519,354 | IssuesEvent | 2021-10-05 17:51:09 | FlukeAndFeather/stickleback | https://api.github.com/repos/FlukeAndFeather/stickleback | opened | GIF links broken in PyPI README | documentation | See https://pypi.org/project/stickleback/0.1.1/, section "Visualize sensor and event data". Use absolute URLs or something instead? | 1.0 | GIF links broken in PyPI README - See https://pypi.org/project/stickleback/0.1.1/, section "Visualize sensor and event data". Use absolute URLs or something instead? | non_infrastructure | gif links broken in pypi readme see section visualize sensor and event data use absolute urls or something instead | 0 |
5,513 | 5,714,936,500 | IssuesEvent | 2017-04-19 11:49:58 | m-labs/artiq | https://api.github.com/repos/m-labs/artiq | closed | conda leaves build files around | area:infrastructure | After a build, files are left over in ``/var/lib/buildbot/slaves/debian-stretch-amd64-2/miniconda/conda-bld``. This wastes disk space, and more importantly randomly breaks builds when the file names conflict, resulting in errors like:
``fatal: destination path '/var/lib/buildbot/slaves/debian-stretch-amd64-2/miniconda/conda-bld/artiq-kc705-nist_clock_1491023827445/work' already exists and is not an empty directory.`` | 1.0 | conda leaves build files around - After a build, files are left over in ``/var/lib/buildbot/slaves/debian-stretch-amd64-2/miniconda/conda-bld``. This wastes disk space, and more importantly randomly breaks builds when the file names conflict, resulting in errors like:
``fatal: destination path '/var/lib/buildbot/slaves/debian-stretch-amd64-2/miniconda/conda-bld/artiq-kc705-nist_clock_1491023827445/work' already exists and is not an empty directory.`` | infrastructure | conda leaves build files around after a build files are left over in var lib buildbot slaves debian stretch miniconda conda bld this wastes disk space and more importantly randomly breaks builds when the file names conflict resulting in errors like fatal destination path var lib buildbot slaves debian stretch miniconda conda bld artiq nist clock work already exists and is not an empty directory | 1 |
14,657 | 10,209,341,898 | IssuesEvent | 2019-08-14 12:29:51 | microsoft/azure-pipelines-tasks | https://api.github.com/repos/microsoft/azure-pipelines-tasks | closed | Add support for multiple deployments in parallel in AzureRmWebAppDeploymentV4 | Area: AzureAppService Area: Release enhancement | ## Required Information
**Question, Bug, or Feature?**
*Type*: Feature
**Enter Task Name**: AzureRmWebAppDeployment
## Environment
- Server - Azure Pipelines or TFS on-premises?
- Azure Pipelines
- Agent - Hosted or Private:
- Private with Windows
## Issue Description
Currently, the AzureRmWebAppDeploymentV4 task allows to deploy to a single Azure App Service. We created a pull request to update it to allow parallel deployments to multiple Azure App Services. Link to the PR: [10909](https://github.com/microsoft/azure-pipelines-tasks/pull/10909)
This pull request adds support for deploying to multiple Azure App Services in one task execution. The services are specified in the App Service name field of the task separated by commas.

The task reads the list of names and deploys the app to the services in parallel.
A pre-release version was tested locally and in Azure Pipelines. The names of the apps were hidden from the picture:

Thank you for your time. | 1.0 | Add support for multiple deployments in parallel in AzureRmWebAppDeploymentV4 - ## Required Information
**Question, Bug, or Feature?**
*Type*: Feature
**Enter Task Name**: AzureRmWebAppDeployment
## Environment
- Server - Azure Pipelines or TFS on-premises?
- Azure Pipelines
- Agent - Hosted or Private:
- Private with Windows
## Issue Description
Currently, the AzureRmWebAppDeploymentV4 task allows to deploy to a single Azure App Service. We created a pull request to update it to allow parallel deployments to multiple Azure App Services. Link to the PR: [10909](https://github.com/microsoft/azure-pipelines-tasks/pull/10909)
This pull request adds support for deploying to multiple Azure App Services in one task execution. The services are specified in the App Service name field of the task separated by commas.

The task reads the list of names and deploys the app to the services in parallel.
A pre-release version was tested locally and in Azure Pipelines. The names of the apps were hidden from the picture:

Thank you for your time. | non_infrastructure | add support for multiple deployments in parallel in required information question bug or feature type feature enter task name azurermwebappdeployment environment server azure pipelines or tfs on premises azure pipelines agent hosted or private private with windows issue description currently the task allows to deploy to a single azure app service we created a pull request to update it to allow parallel deployments to multiple azure app services link to the pr this pull request adds support for deploying to multiple azure app services in one task execution the services are specified in the app service name field of the task separated by commas the task reads the list of names and deploys the app to the services in parallel a pre release version was tested locally and in azure pipelines the names of the apps were hidden from the picture thank you for your time | 0 |
27,298 | 21,579,403,778 | IssuesEvent | 2022-05-02 17:01:30 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | older repository references from the SDK DEPS file | area-infrastructure | There are a few older repository references in our SDK DEPS file. These are repositories that may at one point have lived in the `dart-lang/` github org but have since moved out. Github re-directs references to the new location automatically; I don't know how important it is for the references to be correct (or, if we have to update our mirror information). It would make some tooling I'm writing simpler if the DEPS file points to the current (non-redirected) locations. The references are:
- `https://github.com/google/protobuf.dart` (currently referred to as 'dart-lang/protobuf.dart')
- `https://github.com/lrhn/charcode` (currently referred to as 'dart-lang/charcode')
- `https://github.com/google/pedantic` (currently referred to as 'dart-lang/pedantic')
| 1.0 | older repository references from the SDK DEPS file - There are a few older repository references in our SDK DEPS file. These are repositories that may at one point have lived in the `dart-lang/` github org but have since moved out. Github re-directs references to the new location automatically; I don't know how important it is for the references to be correct (or, if we have to update our mirror information). It would make some tooling I'm writing simpler if the DEPS file points to the current (non-redirected) locations. The references are:
- `https://github.com/google/protobuf.dart` (currently referred to as 'dart-lang/protobuf.dart')
- `https://github.com/lrhn/charcode` (currently referred to as 'dart-lang/charcode')
- `https://github.com/google/pedantic` (currently referred to as 'dart-lang/pedantic')
| infrastructure | older repository references from the sdk deps file there are a few older repository references in our sdk deps file these are repositories that may at one point have lived in the dart lang github org but have since moved out github re directs references to the new location automatically i don t know how important it is for the references to be correct or if we have to update our mirror information it would make some tooling i m writing simpler if the deps file points to the current non redirected locations the references are currently referred to as dart lang protobuf dart currently referred to as dart lang charcode currently referred to as dart lang pedantic | 1 |
1,074 | 3,030,271,947 | IssuesEvent | 2015-08-04 16:37:20 | google/trace-viewer | https://api.github.com/repos/google/trace-viewer | closed | base.requestIdleCallback is not pumped when the tab is backgrounded | Bug Infrastructure P2 | _From [nd...@chromium.org](https://code.google.com/u/102435256078839283966/) on October 31, 2013 01:30:04_
It should be. We should listen to the page visibility ( https://developer.mozilla.org/en-US/docs/Web/Guide/User_experience/Using_the_Page_Visibility_API ) and when we go invisible, and we have pending rafs pendnig, we should pump the idle callbacks manually.
_Original issue: http://code.google.com/p/trace-viewer/issues/detail?id=494_ | 1.0 | base.requestIdleCallback is not pumped when the tab is backgrounded - _From [nd...@chromium.org](https://code.google.com/u/102435256078839283966/) on October 31, 2013 01:30:04_
It should be. We should listen to the page visibility ( https://developer.mozilla.org/en-US/docs/Web/Guide/User_experience/Using_the_Page_Visibility_API ) and when we go invisible, and we have pending rafs pendnig, we should pump the idle callbacks manually.
_Original issue: http://code.google.com/p/trace-viewer/issues/detail?id=494_ | infrastructure | base requestidlecallback is not pumped when the tab is backgrounded from on october it should be we should listen to the page visibility and when we go invisible and we have pending rafs pendnig we should pump the idle callbacks manually original issue | 1 |
16,408 | 11,954,165,197 | IssuesEvent | 2020-04-03 22:43:59 | nwfsc-fram/boatnet | https://api.github.com/repos/nwfsc-fram/boatnet | opened | Move bn-common to boatnet-modules repo | Prj:infrastructure dependencies | Started on this a while ago, but had issues with linking at the time.
Should be OK to move this now, will ease development I believe (need to use yarn link during dev, but won't need to rebuild whole lerna workspace) | 1.0 | Move bn-common to boatnet-modules repo - Started on this a while ago, but had issues with linking at the time.
Should be OK to move this now, will ease development I believe (need to use yarn link during dev, but won't need to rebuild whole lerna workspace) | infrastructure | move bn common to boatnet modules repo started on this a while ago but had issues with linking at the time should be ok to move this now will ease development i believe need to use yarn link during dev but won t need to rebuild whole lerna workspace | 1 |
165,522 | 26,184,850,110 | IssuesEvent | 2023-01-02 21:46:48 | phetsims/calculus-grapher | https://api.github.com/repos/phetsims/calculus-grapher | closed | CueingArrows for f(x) and predict curve | design:general status:ready-for-review | While working on encapsulating the creation of CueingArrow in a separate class (#120), it occurred to me that the behavior of CueingArrows is underspecified when we have both predict curve and f(x) curve.
Trying to think of the interactions, I would suggest that each curve would have cueing arrows "attached" to them. As a user would push on the predict button, the predict curve would appear with cueing arrows. This would clue in the user that the predict Curve is inasmuch interactive as the f(x) curve.
Assigning to @amanda-phet to see if she concurs or has a better idea.
| 1.0 | CueingArrows for f(x) and predict curve - While working on encapsulating the creation of CueingArrow in a separate class (#120), it occurred to me that the behavior of CueingArrows is underspecified when we have both predict curve and f(x) curve.
Trying to think of the interactions, I would suggest that each curve would have cueing arrows "attached" to them. As a user would push on the predict button, the predict curve would appear with cueing arrows. This would clue in the user that the predict Curve is inasmuch interactive as the f(x) curve.
Assigning to @amanda-phet to see if she concurs or has a better idea.
| non_infrastructure | cueingarrows for f x and predict curve while working on encapsulating the creation of cueingarrow in a separate class it occurred to me that the behavior of cueingarrows is underspecified when we have both predict curve and f x curve trying to think of the interactions i would suggest that each curve would have cueing arrows attached to them as a user would push on the predict button the predict curve would appear with cueing arrows this would clue in the user that the predict curve is inasmuch interactive as the f x curve assigning to amanda phet to see if she concurs or has a better idea | 0 |
30,880 | 2,728,807,734 | IssuesEvent | 2015-04-16 01:31:41 | Saveyour-Team/Saveyour | https://api.github.com/repos/Saveyour-Team/Saveyour | opened | When a module does not exist/is not downloaded, yet the user's file has settings for that module, the program will crash | bug priority-medium | I noticed this issue when attempting to pull changes from WeeklyToDo for UI reformatting. I had settings for Google Calendar that were to be loaded; however, since the Google Calendar module did not exist in that branch, the load() method kept crashing from trying to load a nonexistent module. | 1.0 | When a module does not exist/is not downloaded, yet the user's file has settings for that module, the program will crash - I noticed this issue when attempting to pull changes from WeeklyToDo for UI reformatting. I had settings for Google Calendar that were to be loaded; however, since the Google Calendar module did not exist in that branch, the load() method kept crashing from trying to load a nonexistent module. | non_infrastructure | when a module does not exist is not downloaded yet the user s file has settings for that module the program will crash i noticed this issue when attempting to pull changes from weeklytodo for ui reformatting i had settings for google calendar that were to be loaded however since the google calendar module did not exist in that branch the load method kept crashing from trying to load a nonexistent module | 0 |
15,096 | 9,477,326,596 | IssuesEvent | 2019-04-19 18:13:28 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | Microsoft Security Advisory CVE-2019-0564: ASP.NET Core Denial of Service Vulnerability | Security discussion | # Microsoft Security Advisory CVE-2019-0564: ASP.NET Core Denial of Service Vulnerability
## <a name="executive-summary"></a>Executive summary
Microsoft is releasing this security advisory to provide information about a vulnerability in ASP.NET Core 2.1 and 2.2. This advisory also provides guidance on what developers can do to update their applications to remove this vulnerability.
Microsoft is aware of a denial of service vulnerability exists when ASP.NET Core improperly handles web requests. An attacker who successfully exploited this vulnerability could cause a denial of service against an ASP.NET Core web application. The vulnerability can be exploited remotely, without authentication.
A remote unauthenticated attacker could exploit this vulnerability by issuing specially crafted requests to the .NET Core application.
The update addresses the vulnerability by correcting how the ASP.NET Core web application handles web requests.
## Announcement
The original announcement for this issue can be found at https://github.com/aspnet/Announcements/issues/334
### <a name="mitigation-factors"></a>Mitigation factors
Microsoft has not identified any mitigating factors for this vulnerability.
## <a name="affected-software"></a>Affected software
Any .NET Core based application that uses any of following vulnerable packages:
Package name | Vulnerable versions | Secure versions
------------ | ------------------- | -------------------------
Microsoft.AspNetCore.WebSockets | 2.2.0<br/>2.1.0, 2.1.1 | 2.2.1<br/>2.1.7
Microsoft.AspNetCore.Server.Kestrel.Core | 2.1.0, 2.1.1, 2.1.2, 2.1.3 | 2.1.7
System.Net.WebSockets.WebSocketProtocol | 4.5.0, 4.5.1, 4.5.2 | 4.5.3
Microsoft.NETCore.App | 2.2.0<br/>2.1.0, 2.1.1, 2.1.2, 2.1.3, 2.1.4, 2.1.5, 2.1.6 | 2.2.1 <br/> 2.1.7
Microsoft.AspNetCore.App | 2.2.0<br/>2.1.0, 2.1.1, 2.1.2, 2.1.3, 2.1.4, 2.1.5, 2.1.6 | 2.2.1 <br/> 2.1.7
Microsoft.AspNetCore.All | 2.2.0<br/>2.1.0, 2.1.1, 2.1.2, 2.1.3, 2.1.4, 2.1.5, 2.1.6 | 2.2.1 <br/> 2.1.7
## <a name="advisory-faq"></a>Advisory FAQ
### <a name="how-affected"></a>How do I know if I am affected?
#### Applications that use Microsoft.AspNetCore.App or Microsoft.AspNetCore.All
The [vulnerable packages](#affected-software) are distributed as part of the .NET Core 2.1 runtime. To check the currently installed runtimes, open a command prompt and run the `dotnet --info` command. If you have a 2.1 or greater runtime installed, you'll see output like the following;
```text
.NET Core SDK (reflecting any global.json):
Version: 2.2.100
Commit: b9f2fa0ca8
Runtime Environment:
OS Name: Windows
OS Version: 10.0.17134
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\2.2.100\
Host (useful for support):
Version: 2.2.0
Commit: 1249f08fed
.NET Core SDKs installed:
2.1.5 [C:\Program Files\dotnet\sdk]
2.1.502 [C:\Program Files\dotnet\sdk]
2.2.100 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.All 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
If your host version is 2.1.x and the highest `Microsoft.AspNetCore.App` runtime version is less than 2.1.7, you're vulnerable to this issue.
If your host version is 2.2.x and the highest 2.1.x `Microsoft.AspNetCore.App` runtime version is less than 2.1.7, or the highest 2.2.x `Microsoft.AspNetCore.App` runtime version is less than 2.2.1, you're vulnerable to this issue.
You will not see SDK versions if you have only installed the runtime.
#### Applications not using Microsoft.AspNetCore.App or Microsoft.AspNetCore.All
Some of the affected assemblies are also available separately as a NuGet package. If you are not using Microsoft.AspNetCore.App or .All, you may still be affected if you are referencing the packages listed above.
##### Direct dependencies
Direct dependencies are discoverable by examining your `csproj` file. They can be fixed by [editing the project file](#fixing-direct-dependencies) or using NuGet to update the dependency.
##### Transitive dependencies
Transitive dependencies occur when you add a package to your project that in turn relies on another package. For example, if Contoso publishes a package `Contoso.Utility` which, in turn, depends on `Contoso.Internals` and you add the `Contoso.Utility` package to your project now your project has a direct dependency on `Contoso.Utility` and, because `Contoso.Utility` depends 'Contoso.Internals', your application gains a transitive dependency on the `Contoso.Internals` package.
Transitive dependencies are reviewable in two ways:
* In the Visual Studio Solution Explorer window, which supports searching.
* By examining the `project.assets.json` file contained in the obj directory of your project for `csproj` based projects
The `project.assets.json` files are the authoritative list of all packages used by your project, containing both direct and transitive dependencies.
There are two ways to view transitive dependencies. You can either [use Visual Studio’s Solution Explorer](#vs-solution-explorer), or you can review [the `project.assets.json` file](#project-assets-json)).
##### <a name="vs-solution-explorer"></a>Using Visual Studio Solution Explorer
To use Solution Explorer, open the project in Visual Studio, and then press Ctrl+; to activate the search in Solution Explorer. Search for the [vulnerable package](#affected-software) and make a note of the version numbers of any results you find.
For example, searching for `Microsoft.AspNetCore.Mvc.Core` in an example project that contains a package that takes a dependency on `Microsoft.AspNetCore.Mvc` shows the following results in Visual Studio 2017:

The search results appear as a tree. In the previous results, you can see that a reference to `Microsoft.AspNetCore.Mvc.Core` version 1.1.2 is discovered.
Under the Dependencies node is a NuGet node. Under the NuGet node is the list of packages you have directly taken a dependency on and their versions.
In screenshot, the application takes a direct dependency on `Microsoft.AspNetCore.Mvc`. `Microsoft.AspNetCore.Mvc` in turn has leaf nodes that list its dependencies and their versions.
The `Microsoft.AspNetCore.Mvc` package takes a dependency on a version of `Microsoft.AspNetCore.Mvc.ApiExplorer`, that in turn takes a dependency on a version of `Microsoft.AspNetCore.Mvc.Core`.
##### <a name="project-assets-json"></a> Manually reviewing project.assets.json
Open the *project.assets.json* file from your project’s obj directory in your editor. We suggest you use an editor that understands JSON and allows you to collapse and expand nodes to review this file.
Visual Studio and Visual Studio Code provide JSON friendly editing.
Search the *project.assets.json* file for the [vulnerable package](#affected-software), using the format `packagename/` for each of the package names from the preceding table. If you find the assembly name in your search:
* Examine the line on which they are found, the version number is after the `/`.
* Compare to the [vulnerable versions table](#affected-software).
For example, a search result that shows `System.IO.Pipelines/4.5.0` is a reference to version 4.5.0 of `System.IO.Pipelines`.
If your *project.assets.json* file includes references to the [vulnerable package](#affected-software), then you need to fix the transitive dependencies.
If you have not found any reference to any vulnerable packages, this means either
* None of your direct dependencies depend on any vulnerable packages, or
* You have already fixed the problem by updating the direct dependencies.
### <a name="how-fix"></a>How do I fix the issue?
#### Updating the version of Microsoft.NETCore.App, Microsoft.AspNetCore.App, and Microsoft.AspNetCore.All
If your host version is a 2.1.x host, you must install version 2.1.7 or later of the [.NET Core 2.1 runtime](https://dotnet.microsoft.com/download/dotnet-core/2.1), or corresponding [.NET Core 2.1 SDK](https://dotnet.microsoft.com/download/dotnet-core/2.1).
If your host version is a 2.2.x, you must install version 2.2.1 or later of the [.NET Core 2.2 runtime](https://dotnet.microsoft.com/download/dotnet-core/2.2), or corresponding [.NET Core 2.2 SDK](https://dotnet.microsoft.com/download/dotnet-core/2.2).
#### <a name="fixing-direct-dependencies"></a>Fixing direct dependencies
Open *projectname.csproj* in your editor. If you're using Visual Studio, right-click the project and choose **Edit projectname.csproj** from the context menu, where projectname is the name of your project. Look for `PackageReference` elements. The following shows an example project file:
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<TargetFramework>netcoreapp2.1</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="System.IO.Pipelines" Version="4.5.0" />
</ItemGroup>
</Project>
```
The preceding example has a reference to the [vulnerable package](#affected-software), as seen by the single `PackageReference` element. The name of the package is in the `Include` attribute.
The package version number is in the `Version` attribute. The previous example shows a single direct dependency on `System.IO.Pipelines` version 4.5.0
To update the version to the secure package, change the version number to the updated package version as listed on the table [previously](#affected-software).
In this example, update `System.IO.Pipelines` to the appropriate [fixed package number](#affected-software) for your major version. Save the *csproj* file. The example *csproj* now looks as follows:
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<TargetFramework>netcoreapp2.1</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="System.IO.Pipelines" Version="4.5.1" />
</ItemGroup>
</Project>
```
If you're using Visual Studio and you save your updated *csproj* file, Visual Studio will restore the new package version.
You can see the restore results by opening the **Output** window (Ctrl+Alt+O) and changing the **Show output from** drop-down list to **Package Manager**.
If you're not using Visual Studio, open a command line and change to your project directory. Execute the `dotnet restore` command to restore the updated dependencies.
Now recompile your application. If after recompilation you see a *Dependency conflict warning*, you must update your other direct dependencies to versions that take a dependency on the updated package.
#### <a name="transitive-dependencies"></a>Fixing transitive dependencies
If your transitive dependency review found references to the [vulnerable package](#affected-software), you must add a direct dependency to the updated package to your *csproj* file to override the transitive dependency.
Open *projectname.csproj* in your editor. If you're using Visual Studio, right-click the project and choose **Edit projectname.csproj** from the context menu, where projectname is the name of your project.
Look for `PackageReference` nodes, for example:
```xml
<Project Sdk="Microsoft.NET.Sdk.">
<PropertyGroup>
<TargetFramework>net461</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="ThirdParty.NotUpdatedYet" Version="2.0.0" />
</ItemGroup>
</Project>
```
You must add a direct dependency to the updated version of the [vulnerable package](#affected-software) by adding it to the *csproj* file.
You do this by adding a new line to the dependencies section, referencing the fixed version.
For example, if your search showed a transitive reference to a vulnerable `System.ServiceModel.Http` version, you'd add a reference to the [fixed package number](#affected-software).
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<TargetFramework>net461</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="System.IO.Pipelines" Version="4.5.1" />
<PackageReference Include="ThirdParty.NotUpdatedYet" Version="2.0.0" />
</ItemGroup>
</Project>
```
After you've added the direct dependency reference, save your *csproj* file.
If you're using Visual Studio, save your updated *csproj* file and Visual Studio will restore the new package versions.
You can see the restore results by opening the **Output** window (Ctrl+Alt+O) and changing the **Show output from** drop-down list to **Package Manager**.
If you're not using Visual Studio, open a command line and change to your project directory. Execute the `dotnet restore` command to restore the new dependencies.
##### Rebuilding your application
Finally, you must rebuild your application, test, and redeploy.
## Other Information
### Reporting Security Issues
If you have found a potential security issue in .NET Core, please email details to secure@microsoft.com. Reports may qualify for the .NET Core Bug Bounty. Details of the .NET Core Bug Bounty including terms and conditions are at [https://aka.ms/corebounty](https://aka.ms/corebounty).
### Support
You can ask questions about this issue on GitHub in the .NET Core or ASP.NET Core organizations. These are located at https://github.com/dotnet/ and https://github.com/aspnet/. The Announcements repo for each product (https://github.com/dotnet/Announcements and https://github.com/aspnet/Announcements) will contain this bulletin as an issue and will include a link to a discussion issue. You can ask questions in the discussion issue.
### Disclaimer
The information provided in this advisory is provided "as is" without warranty of any kind. Microsoft disclaims all warranties, either express or implied, including the warranties of merchantability and fitness for a particular purpose. In no event shall Microsoft Corporation or its suppliers be liable for any damages whatsoever including direct, indirect, incidental, consequential, loss of business profits or special damages, even if Microsoft Corporation or its suppliers have been advised of the possibility of such damages. Some states do not allow the exclusion or limitation of liability for consequential or incidental damages so the foregoing limitation may not apply.
### External Links
[CVE-2019-0564](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0564)
### Revisions
V1.0 (January 8, 2019): Advisory published.
_Version 1.0_
_Last Updated 2019-01-08_
| True | Microsoft Security Advisory CVE-2019-0564: ASP.NET Core Denial of Service Vulnerability - # Microsoft Security Advisory CVE-2019-0564: ASP.NET Core Denial of Service Vulnerability
## <a name="executive-summary"></a>Executive summary
Microsoft is releasing this security advisory to provide information about a vulnerability in ASP.NET Core 2.1 and 2.2. This advisory also provides guidance on what developers can do to update their applications to remove this vulnerability.
Microsoft is aware of a denial of service vulnerability exists when ASP.NET Core improperly handles web requests. An attacker who successfully exploited this vulnerability could cause a denial of service against an ASP.NET Core web application. The vulnerability can be exploited remotely, without authentication.
A remote unauthenticated attacker could exploit this vulnerability by issuing specially crafted requests to the .NET Core application.
The update addresses the vulnerability by correcting how the ASP.NET Core web application handles web requests.
## Announcement
The original announcement for this issue can be found at https://github.com/aspnet/Announcements/issues/334
### <a name="mitigation-factors"></a>Mitigation factors
Microsoft has not identified any mitigating factors for this vulnerability.
## <a name="affected-software"></a>Affected software
Any .NET Core based application that uses any of following vulnerable packages:
Package name | Vulnerable versions | Secure versions
------------ | ------------------- | -------------------------
Microsoft.AspNetCore.WebSockets | 2.2.0<br/>2.1.0, 2.1.1 | 2.2.1<br/>2.1.7
Microsoft.AspNetCore.Server.Kestrel.Core | 2.1.0, 2.1.1, 2.1.2, 2.1.3 | 2.1.7
System.Net.WebSockets.WebSocketProtocol | 4.5.0, 4.5.1, 4.5.2 | 4.5.3
Microsoft.NETCore.App | 2.2.0<br/>2.1.0, 2.1.1, 2.1.2, 2.1.3, 2.1.4, 2.1.5, 2.1.6 | 2.2.1 <br/> 2.1.7
Microsoft.AspNetCore.App | 2.2.0<br/>2.1.0, 2.1.1, 2.1.2, 2.1.3, 2.1.4, 2.1.5, 2.1.6 | 2.2.1 <br/> 2.1.7
Microsoft.AspNetCore.All | 2.2.0<br/>2.1.0, 2.1.1, 2.1.2, 2.1.3, 2.1.4, 2.1.5, 2.1.6 | 2.2.1 <br/> 2.1.7
## <a name="advisory-faq"></a>Advisory FAQ
### <a name="how-affected"></a>How do I know if I am affected?
#### Applications that use Microsoft.AspNetCore.App or Microsoft.AspNetCore.All
The [vulnerable packages](#affected-software) are distributed as part of the .NET Core 2.1 runtime. To check the currently installed runtimes, open a command prompt and run the `dotnet --info` command. If you have a 2.1 or greater runtime installed, you'll see output like the following;
```text
.NET Core SDK (reflecting any global.json):
Version: 2.2.100
Commit: b9f2fa0ca8
Runtime Environment:
OS Name: Windows
OS Version: 10.0.17134
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\2.2.100\
Host (useful for support):
Version: 2.2.0
Commit: 1249f08fed
.NET Core SDKs installed:
2.1.5 [C:\Program Files\dotnet\sdk]
2.1.502 [C:\Program Files\dotnet\sdk]
2.2.100 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.All 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
If your host version is 2.1.x and the highest `Microsoft.AspNetCore.App` runtime version is less than 2.1.7, you're vulnerable to this issue.
If your host version is 2.2.x and the highest 2.1.x `Microsoft.AspNetCore.App` runtime version is less than 2.1.7, or the highest 2.2.x `Microsoft.AspNetCore.App` runtime version is less than 2.2.1, you're vulnerable to this issue.
You will not see SDK versions if you have only installed the runtime.
#### Applications not using Microsoft.AspNetCore.App or Microsoft.AspNetCore.All
Some of the affected assemblies are also available separately as a NuGet package. If you are not using Microsoft.AspNetCore.App or .All, you may still be affected if you are referencing the packages listed above.
##### Direct dependencies
Direct dependencies are discoverable by examining your `csproj` file. They can be fixed by [editing the project file](#fixing-direct-dependencies) or using NuGet to update the dependency.
##### Transitive dependencies
Transitive dependencies occur when you add a package to your project that in turn relies on another package. For example, if Contoso publishes a package `Contoso.Utility` which, in turn, depends on `Contoso.Internals` and you add the `Contoso.Utility` package to your project now your project has a direct dependency on `Contoso.Utility` and, because `Contoso.Utility` depends 'Contoso.Internals', your application gains a transitive dependency on the `Contoso.Internals` package.
Transitive dependencies are reviewable in two ways:
* In the Visual Studio Solution Explorer window, which supports searching.
* By examining the `project.assets.json` file contained in the obj directory of your project for `csproj` based projects
The `project.assets.json` files are the authoritative list of all packages used by your project, containing both direct and transitive dependencies.
There are two ways to view transitive dependencies. You can either [use Visual Studio’s Solution Explorer](#vs-solution-explorer), or you can review [the `project.assets.json` file](#project-assets-json)).
##### <a name="vs-solution-explorer"></a>Using Visual Studio Solution Explorer
To use Solution Explorer, open the project in Visual Studio, and then press Ctrl+; to activate the search in Solution Explorer. Search for the [vulnerable package](#affected-software) and make a note of the version numbers of any results you find.
For example, searching for `Microsoft.AspNetCore.Mvc.Core` in an example project that contains a package that takes a dependency on `Microsoft.AspNetCore.Mvc` shows the following results in Visual Studio 2017:

The search results appear as a tree. In the previous results, you can see that a reference to `Microsoft.AspNetCore.Mvc.Core` version 1.1.2 is discovered.
Under the Dependencies node is a NuGet node. Under the NuGet node is the list of packages you have directly taken a dependency on and their versions.
In screenshot, the application takes a direct dependency on `Microsoft.AspNetCore.Mvc`. `Microsoft.AspNetCore.Mvc` in turn has leaf nodes that list its dependencies and their versions.
The `Microsoft.AspNetCore.Mvc` package takes a dependency on a version of `Microsoft.AspNetCore.Mvc.ApiExplorer`, that in turn takes a dependency on a version of `Microsoft.AspNetCore.Mvc.Core`.
##### <a name="project-assets-json"></a> Manually reviewing project.assets.json
Open the *project.assets.json* file from your project’s obj directory in your editor. We suggest you use an editor that understands JSON and allows you to collapse and expand nodes to review this file.
Visual Studio and Visual Studio Code provide JSON friendly editing.
Search the *project.assets.json* file for the [vulnerable package](#affected-software), using the format `packagename/` for each of the package names from the preceding table. If you find the assembly name in your search:
* Examine the line on which they are found, the version number is after the `/`.
* Compare to the [vulnerable versions table](#affected-software).
For example, a search result that shows `System.IO.Pipelines/4.5.0` is a reference to version 4.5.0 of `System.IO.Pipelines`.
If your *project.assets.json* file includes references to the [vulnerable package](#affected-software), then you need to fix the transitive dependencies.
If you have not found any reference to any vulnerable packages, this means either
* None of your direct dependencies depend on any vulnerable packages, or
* You have already fixed the problem by updating the direct dependencies.
### <a name="how-fix"></a>How do I fix the issue?
#### Updating the version of Microsoft.NETCore.App, Microsoft.AspNetCore.App, and Microsoft.AspNetCore.All
If your host version is a 2.1.x host, you must install version 2.1.7 or later of the [.NET Core 2.1 runtime](https://dotnet.microsoft.com/download/dotnet-core/2.1), or corresponding [.NET Core 2.1 SDK](https://dotnet.microsoft.com/download/dotnet-core/2.1).
If your host version is a 2.2.x, you must install version 2.2.1 or later of the [.NET Core 2.2 runtime](https://dotnet.microsoft.com/download/dotnet-core/2.2), or corresponding [.NET Core 2.2 SDK](https://dotnet.microsoft.com/download/dotnet-core/2.2).
#### <a name="fixing-direct-dependencies"></a>Fixing direct dependencies
Open *projectname.csproj* in your editor. If you're using Visual Studio, right-click the project and choose **Edit projectname.csproj** from the context menu, where projectname is the name of your project. Look for `PackageReference` elements. The following shows an example project file:
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<TargetFramework>netcoreapp2.1</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="System.IO.Pipelines" Version="4.5.0" />
</ItemGroup>
</Project>
```
The preceding example has a reference to the [vulnerable package](#affected-software), as seen by the single `PackageReference` element. The name of the package is in the `Include` attribute.
The package version number is in the `Version` attribute. The previous example shows a single direct dependency on `System.IO.Pipelines` version 4.5.0
To update the version to the secure package, change the version number to the updated package version as listed on the table [previously](#affected-software).
In this example, update `System.IO.Pipelines` to the appropriate [fixed package number](#affected-software) for your major version. Save the *csproj* file. The example *csproj* now looks as follows:
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<TargetFramework>netcoreapp2.1</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="System.IO.Pipelines" Version="4.5.1" />
</ItemGroup>
</Project>
```
If you're using Visual Studio and you save your updated *csproj* file, Visual Studio will restore the new package version.
You can see the restore results by opening the **Output** window (Ctrl+Alt+O) and changing the **Show output from** drop-down list to **Package Manager**.
If you're not using Visual Studio, open a command line and change to your project directory. Execute the `dotnet restore` command to restore the updated dependencies.
Now recompile your application. If after recompilation you see a *Dependency conflict warning*, you must update your other direct dependencies to versions that take a dependency on the updated package.
#### <a name="transitive-dependencies"></a>Fixing transitive dependencies
If your transitive dependency review found references to the [vulnerable package](#affected-software), you must add a direct dependency to the updated package to your *csproj* file to override the transitive dependency.
Open *projectname.csproj* in your editor. If you're using Visual Studio, right-click the project and choose **Edit projectname.csproj** from the context menu, where projectname is the name of your project.
Look for `PackageReference` nodes, for example:
```xml
<Project Sdk="Microsoft.NET.Sdk.">
<PropertyGroup>
<TargetFramework>net461</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="ThirdParty.NotUpdatedYet" Version="2.0.0" />
</ItemGroup>
</Project>
```
You must add a direct dependency to the updated version of the [vulnerable package](#affected-software) by adding it to the *csproj* file.
You do this by adding a new line to the dependencies section, referencing the fixed version.
For example, if your search showed a transitive reference to a vulnerable `System.ServiceModel.Http` version, you'd add a reference to the [fixed package number](#affected-software).
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<TargetFramework>net461</TargetFramework>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="System.IO.Pipelines" Version="4.5.1" />
<PackageReference Include="ThirdParty.NotUpdatedYet" Version="2.0.0" />
</ItemGroup>
</Project>
```
After you've added the direct dependency reference, save your *csproj* file.
If you're using Visual Studio, save your updated *csproj* file and Visual Studio will restore the new package versions.
You can see the restore results by opening the **Output** window (Ctrl+Alt+O) and changing the **Show output from** drop-down list to **Package Manager**.
If you're not using Visual Studio, open a command line and change to your project directory. Execute the `dotnet restore` command to restore the new dependencies.
##### Rebuilding your application
Finally, you must rebuild your application, test, and redeploy.
## Other Information
### Reporting Security Issues
If you have found a potential security issue in .NET Core, please email details to secure@microsoft.com. Reports may qualify for the .NET Core Bug Bounty. Details of the .NET Core Bug Bounty including terms and conditions are at [https://aka.ms/corebounty](https://aka.ms/corebounty).
### Support
You can ask questions about this issue on GitHub in the .NET Core or ASP.NET Core organizations. These are located at https://github.com/dotnet/ and https://github.com/aspnet/. The Announcements repo for each product (https://github.com/dotnet/Announcements and https://github.com/aspnet/Announcements) will contain this bulletin as an issue and will include a link to a discussion issue. You can ask questions in the discussion issue.
### Disclaimer
The information provided in this advisory is provided "as is" without warranty of any kind. Microsoft disclaims all warranties, either express or implied, including the warranties of merchantability and fitness for a particular purpose. In no event shall Microsoft Corporation or its suppliers be liable for any damages whatsoever including direct, indirect, incidental, consequential, loss of business profits or special damages, even if Microsoft Corporation or its suppliers have been advised of the possibility of such damages. Some states do not allow the exclusion or limitation of liability for consequential or incidental damages so the foregoing limitation may not apply.
### External Links
[CVE-2019-0564](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0564)
### Revisions
V1.0 (January 8, 2019): Advisory published.
_Version 1.0_
_Last Updated 2019-01-08_
| non_infrastructure | microsoft security advisory cve asp net core denial of service vulnerability microsoft security advisory cve asp net core denial of service vulnerability executive summary microsoft is releasing this security advisory to provide information about a vulnerability in asp net core and this advisory also provides guidance on what developers can do to update their applications to remove this vulnerability microsoft is aware of a denial of service vulnerability exists when asp net core improperly handles web requests an attacker who successfully exploited this vulnerability could cause a denial of service against an asp net core web application the vulnerability can be exploited remotely without authentication a remote unauthenticated attacker could exploit this vulnerability by issuing specially crafted requests to the net core application the update addresses the vulnerability by correcting how the asp net core web application handles web requests announcement the original announcement for this issue can be found at mitigation factors microsoft has not identified any mitigating factors for this vulnerability affected software any net core based application that uses any of following vulnerable packages package name vulnerable versions secure versions microsoft aspnetcore websockets microsoft aspnetcore server kestrel core system net websockets websocketprotocol microsoft netcore app microsoft aspnetcore app microsoft aspnetcore all advisory faq how do i know if i am affected applications that use microsoft aspnetcore app or microsoft aspnetcore all the affected software are distributed as part of the net core runtime to check the currently installed runtimes open a command prompt and run the dotnet info command if you have a or greater runtime installed you ll see output like the following text net core sdk reflecting any global json version commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk host useful for support version commit net core sdks installed net core runtimes installed microsoft aspnetcore all microsoft aspnetcore app microsoft aspnetcore all microsoft aspnetcore app microsoft netcore app microsoft netcore app to install additional net core runtimes or sdks if your host version is x and the highest microsoft aspnetcore app runtime version is less than you re vulnerable to this issue if your host version is x and the highest x microsoft aspnetcore app runtime version is less than or the highest x microsoft aspnetcore app runtime version is less than you re vulnerable to this issue you will not see sdk versions if you have only installed the runtime applications not using microsoft aspnetcore app or microsoft aspnetcore all some of the affected assemblies are also available separately as a nuget package if you are not using microsoft aspnetcore app or all you may still be affected if you are referencing the packages listed above direct dependencies direct dependencies are discoverable by examining your csproj file they can be fixed by fixing direct dependencies or using nuget to update the dependency transitive dependencies transitive dependencies occur when you add a package to your project that in turn relies on another package for example if contoso publishes a package contoso utility which in turn depends on contoso internals and you add the contoso utility package to your project now your project has a direct dependency on contoso utility and because contoso utility depends contoso internals your application gains a transitive dependency on the contoso internals package transitive dependencies are reviewable in two ways in the visual studio solution explorer window which supports searching by examining the project assets json file contained in the obj directory of your project for csproj based projects the project assets json files are the authoritative list of all packages used by your project containing both direct and transitive dependencies there are two ways to view transitive dependencies you can either vs solution explorer or you can review project assets json using visual studio solution explorer to use solution explorer open the project in visual studio and then press ctrl to activate the search in solution explorer search for the affected software and make a note of the version numbers of any results you find for example searching for microsoft aspnetcore mvc core in an example project that contains a package that takes a dependency on microsoft aspnetcore mvc shows the following results in visual studio the search results appear as a tree in the previous results you can see that a reference to microsoft aspnetcore mvc core version is discovered under the dependencies node is a nuget node under the nuget node is the list of packages you have directly taken a dependency on and their versions in screenshot the application takes a direct dependency on microsoft aspnetcore mvc microsoft aspnetcore mvc in turn has leaf nodes that list its dependencies and their versions the microsoft aspnetcore mvc package takes a dependency on a version of microsoft aspnetcore mvc apiexplorer that in turn takes a dependency on a version of microsoft aspnetcore mvc core manually reviewing project assets json open the project assets json file from your project’s obj directory in your editor we suggest you use an editor that understands json and allows you to collapse and expand nodes to review this file visual studio and visual studio code provide json friendly editing search the project assets json file for the affected software using the format packagename for each of the package names from the preceding table if you find the assembly name in your search examine the line on which they are found the version number is after the compare to the affected software for example a search result that shows system io pipelines is a reference to version of system io pipelines if your project assets json file includes references to the affected software then you need to fix the transitive dependencies if you have not found any reference to any vulnerable packages this means either none of your direct dependencies depend on any vulnerable packages or you have already fixed the problem by updating the direct dependencies how do i fix the issue updating the version of microsoft netcore app microsoft aspnetcore app and microsoft aspnetcore all if your host version is a x host you must install version or later of the or corresponding if your host version is a x you must install version or later of the or corresponding fixing direct dependencies open projectname csproj in your editor if you re using visual studio right click the project and choose edit projectname csproj from the context menu where projectname is the name of your project look for packagereference elements the following shows an example project file xml the preceding example has a reference to the affected software as seen by the single packagereference element the name of the package is in the include attribute the package version number is in the version attribute the previous example shows a single direct dependency on system io pipelines version to update the version to the secure package change the version number to the updated package version as listed on the table affected software in this example update system io pipelines to the appropriate affected software for your major version save the csproj file the example csproj now looks as follows xml if you re using visual studio and you save your updated csproj file visual studio will restore the new package version you can see the restore results by opening the output window ctrl alt o and changing the show output from drop down list to package manager if you re not using visual studio open a command line and change to your project directory execute the dotnet restore command to restore the updated dependencies now recompile your application if after recompilation you see a dependency conflict warning you must update your other direct dependencies to versions that take a dependency on the updated package fixing transitive dependencies if your transitive dependency review found references to the affected software you must add a direct dependency to the updated package to your csproj file to override the transitive dependency open projectname csproj in your editor if you re using visual studio right click the project and choose edit projectname csproj from the context menu where projectname is the name of your project look for packagereference nodes for example xml you must add a direct dependency to the updated version of the affected software by adding it to the csproj file you do this by adding a new line to the dependencies section referencing the fixed version for example if your search showed a transitive reference to a vulnerable system servicemodel http version you d add a reference to the affected software xml after you ve added the direct dependency reference save your csproj file if you re using visual studio save your updated csproj file and visual studio will restore the new package versions you can see the restore results by opening the output window ctrl alt o and changing the show output from drop down list to package manager if you re not using visual studio open a command line and change to your project directory execute the dotnet restore command to restore the new dependencies rebuilding your application finally you must rebuild your application test and redeploy other information reporting security issues if you have found a potential security issue in net core please email details to secure microsoft com reports may qualify for the net core bug bounty details of the net core bug bounty including terms and conditions are at support you can ask questions about this issue on github in the net core or asp net core organizations these are located at and the announcements repo for each product and will contain this bulletin as an issue and will include a link to a discussion issue you can ask questions in the discussion issue disclaimer the information provided in this advisory is provided as is without warranty of any kind microsoft disclaims all warranties either express or implied including the warranties of merchantability and fitness for a particular purpose in no event shall microsoft corporation or its suppliers be liable for any damages whatsoever including direct indirect incidental consequential loss of business profits or special damages even if microsoft corporation or its suppliers have been advised of the possibility of such damages some states do not allow the exclusion or limitation of liability for consequential or incidental damages so the foregoing limitation may not apply external links revisions january advisory published version last updated | 0 |
217,239 | 16,848,840,406 | IssuesEvent | 2021-06-20 04:14:32 | hakehuang/infoflow | https://api.github.com/repos/hakehuang/infoflow | opened |
tests-ci :kernel.memory_protection.inherit_resource_pool : zephyr-v2.6.0-286-g46029914a7ac: lpcxpresso55s28: test Timeout
| area: Tests |
**Describe the bug**
kernel.memory_protection.inherit_resource_pool test is Timeout on zephyr-v2.6.0-286-g46029914a7ac on lpcxpresso55s28
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p lpcxpresso55s28 --testcase-root tests --sub-test kernel.memory_protection
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
-
*** Booting Zephyr OS build zephyr-v2.6.0-286-g46029914a7ac ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| 1.0 |
tests-ci :kernel.memory_protection.inherit_resource_pool : zephyr-v2.6.0-286-g46029914a7ac: lpcxpresso55s28: test Timeout
-
**Describe the bug**
kernel.memory_protection.inherit_resource_pool test is Timeout on zephyr-v2.6.0-286-g46029914a7ac on lpcxpresso55s28
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p lpcxpresso55s28 --testcase-root tests --sub-test kernel.memory_protection
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
-
*** Booting Zephyr OS build zephyr-v2.6.0-286-g46029914a7ac ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| non_infrastructure | tests ci kernel memory protection inherit resource pool zephyr test timeout describe the bug kernel memory protection inherit resource pool test is timeout on zephyr on see logs for details to reproduce scripts twister device testing device serial dev p testcase root tests sub test kernel memory protection see error expected behavior test pass impact logs and console output booting zephyr os build zephyr running test suite memory protection test suite start test permission inheritance assertion fail west topdir zephyr arch arm core cortex m fault c esf could not be retrieved successfully shall never occur assertion fail west topdir zephyr arch arm core cortex m fault c esf could not be retrieved successfully shall never occur environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr | 0 |
9,730 | 8,128,996,844 | IssuesEvent | 2018-08-17 13:48:00 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Failing to run AOT tests | area-Infrastructure bug os-windows-uwp | Following the instructions in the documentation to run
```
msbuild /t:BuildAndTest /p:TestTFM=netcore50aot /p:TestNugetRuntimeId=win10-x64-aot /p:UseDotNetNativeToolchain=true
```
results in
```
ILC error ILC1104: Could not find file 'corefx\bin\tests\AnyOS.AnyCPU.Debug\System.Linq.Tests\default.netcore50aot\AppxManifest.xml'.
``` | 1.0 | Failing to run AOT tests - Following the instructions in the documentation to run
```
msbuild /t:BuildAndTest /p:TestTFM=netcore50aot /p:TestNugetRuntimeId=win10-x64-aot /p:UseDotNetNativeToolchain=true
```
results in
```
ILC error ILC1104: Could not find file 'corefx\bin\tests\AnyOS.AnyCPU.Debug\System.Linq.Tests\default.netcore50aot\AppxManifest.xml'.
``` | infrastructure | failing to run aot tests following the instructions in the documentation to run msbuild t buildandtest p testtfm p testnugetruntimeid aot p usedotnetnativetoolchain true results in ilc error could not find file corefx bin tests anyos anycpu debug system linq tests default appxmanifest xml | 1 |
260,617 | 8,212,519,408 | IssuesEvent | 2018-09-04 16:34:10 | cceh/capitularia | https://api.github.com/repos/cceh/capitularia | closed | Ersetzungen von Text in listBibl type="lit", "cat", "abb" | High Priority needs edit freeze | @Capitularia-sk @domtru @skheinen
In den bibliographischen Angaben der Hss.-Dateien (teiHeader/msDesc/listBibl type="lit" o. "cat" o. "abb") sollen einige Einträge durch die Kurztitel ersetzt werden, die wir in unserer Bibliographie verwenden.
Die Liste mit den Einträgen und den entsprechenden Ersetzungen hänge ich an (wegen der Klämmerchen).
[Kurztitel_ersetzen_Liste.docx](https://github.com/cceh/capitularia/files/2005474/Kurztitel_ersetzen_Liste.docx)
| 1.0 | Ersetzungen von Text in listBibl type="lit", "cat", "abb" - @Capitularia-sk @domtru @skheinen
In den bibliographischen Angaben der Hss.-Dateien (teiHeader/msDesc/listBibl type="lit" o. "cat" o. "abb") sollen einige Einträge durch die Kurztitel ersetzt werden, die wir in unserer Bibliographie verwenden.
Die Liste mit den Einträgen und den entsprechenden Ersetzungen hänge ich an (wegen der Klämmerchen).
[Kurztitel_ersetzen_Liste.docx](https://github.com/cceh/capitularia/files/2005474/Kurztitel_ersetzen_Liste.docx)
| non_infrastructure | ersetzungen von text in listbibl type lit cat abb capitularia sk domtru skheinen in den bibliographischen angaben der hss dateien teiheader msdesc listbibl type lit o cat o abb sollen einige einträge durch die kurztitel ersetzt werden die wir in unserer bibliographie verwenden die liste mit den einträgen und den entsprechenden ersetzungen hänge ich an wegen der klämmerchen | 0 |
798,058 | 28,214,847,964 | IssuesEvent | 2023-04-05 08:11:52 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [CDCSDK] Connector support for Colocated Snapshot | kind/enhancement priority/medium area/cdcsdk | Jira Link: [DB-5799](https://yugabyte.atlassian.net/browse/DB-5799)
The support for server side changes were being tracked through: https://github.com/yugabyte/yugabyte-db/issues/13981
Need to provide corresponding connector side support for colocated snapshots.
[DB-5799]: https://yugabyte.atlassian.net/browse/DB-5799?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [CDCSDK] Connector support for Colocated Snapshot - Jira Link: [DB-5799](https://yugabyte.atlassian.net/browse/DB-5799)
The support for server side changes were being tracked through: https://github.com/yugabyte/yugabyte-db/issues/13981
Need to provide corresponding connector side support for colocated snapshots.
[DB-5799]: https://yugabyte.atlassian.net/browse/DB-5799?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | non_infrastructure | connector support for colocated snapshot jira link the support for server side changes were being tracked through need to provide corresponding connector side support for colocated snapshots | 0 |
13,900 | 10,532,635,200 | IssuesEvent | 2019-10-01 11:13:51 | elastic/beats | https://api.github.com/repos/elastic/beats | opened | [Metricbeat] Tomcat module | :infrastructure Metricbeat | # Metricbeat Module / Dataset release checklist
This checklist is intended for Devs which create or update a module to make sure modules are consistent.
## Modules
For a metricset to go GA, the following criterias should be met:
* [ ] Supported versions are documented
* [ ] Supported operating systems are documented (if applicable)
* [ ] Integration tests exist
* [ ] System tests exist
* [ ] Automated checks that all fields are documented
* [ ] Documentation
* [ ] Fields follow [ECS](https://github.com/elastic/ecs) and [naming conventions](https://www.elastic.co/guide/en/beats/devguide/master/event-conventions.html)
* [ ] Dashboards exists (if applicable)
* [ ] Kibana Home Tutorial (if applicable)
* [ ] Open issue in [EUI repo](https://github.com/elastic/eui) to add [icon for module](https://elastic.github.io/eui/#/display/icons) if not already exists.
* [ ] Open PR against Kibana repo with tutorial. Examples can be found [here](https://github.com/elastic/kibana/tree/master/src/legacy/core_plugins/kibana/server/tutorials).
## Metricbeat module
* [ ] Example `data.json` exists and an automated way to generate it exists (`go test -data`)
* [ ] Test environment in Docker exist for integration tests
| 1.0 | [Metricbeat] Tomcat module - # Metricbeat Module / Dataset release checklist
This checklist is intended for Devs which create or update a module to make sure modules are consistent.
## Modules
For a metricset to go GA, the following criterias should be met:
* [ ] Supported versions are documented
* [ ] Supported operating systems are documented (if applicable)
* [ ] Integration tests exist
* [ ] System tests exist
* [ ] Automated checks that all fields are documented
* [ ] Documentation
* [ ] Fields follow [ECS](https://github.com/elastic/ecs) and [naming conventions](https://www.elastic.co/guide/en/beats/devguide/master/event-conventions.html)
* [ ] Dashboards exists (if applicable)
* [ ] Kibana Home Tutorial (if applicable)
* [ ] Open issue in [EUI repo](https://github.com/elastic/eui) to add [icon for module](https://elastic.github.io/eui/#/display/icons) if not already exists.
* [ ] Open PR against Kibana repo with tutorial. Examples can be found [here](https://github.com/elastic/kibana/tree/master/src/legacy/core_plugins/kibana/server/tutorials).
## Metricbeat module
* [ ] Example `data.json` exists and an automated way to generate it exists (`go test -data`)
* [ ] Test environment in Docker exist for integration tests
| infrastructure | tomcat module metricbeat module dataset release checklist this checklist is intended for devs which create or update a module to make sure modules are consistent modules for a metricset to go ga the following criterias should be met supported versions are documented supported operating systems are documented if applicable integration tests exist system tests exist automated checks that all fields are documented documentation fields follow and dashboards exists if applicable kibana home tutorial if applicable open issue in to add if not already exists open pr against kibana repo with tutorial examples can be found metricbeat module example data json exists and an automated way to generate it exists go test data test environment in docker exist for integration tests | 1 |
6,584 | 6,523,489,767 | IssuesEvent | 2017-08-29 08:53:51 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Two buildbots attempt to run test.py with unrecognized '--failure-summary' | area-infrastructure | Somehow two sharded buildbots (out of 14) try to run tests with ''--failure-summary", which fail because of that:
- https://build.chromium.org/p/client.dart.fyi/builders/vm-kernel-precomp-linux-release-x64-5-6-be/builds/2398/steps/vm%20tests/logs/stdio
- https://build.chromium.org/p/client.dart.fyi/builders/vm-kernel-precomp-linux-debug-x64-1-8-be/builds/2253/steps/vm%20tests/logs/stdio
Compare to https://build.chromium.org/p/client.dart.fyi/builders/vm-kernel-precomp-linux-release-x64-4-6-be/builds/1282/steps/vm%20tests/logs/stdio, for example.
@whesse , is this something you can help with or can give pointers where the the config files that should be fixed please? | 1.0 | Two buildbots attempt to run test.py with unrecognized '--failure-summary' - Somehow two sharded buildbots (out of 14) try to run tests with ''--failure-summary", which fail because of that:
- https://build.chromium.org/p/client.dart.fyi/builders/vm-kernel-precomp-linux-release-x64-5-6-be/builds/2398/steps/vm%20tests/logs/stdio
- https://build.chromium.org/p/client.dart.fyi/builders/vm-kernel-precomp-linux-debug-x64-1-8-be/builds/2253/steps/vm%20tests/logs/stdio
Compare to https://build.chromium.org/p/client.dart.fyi/builders/vm-kernel-precomp-linux-release-x64-4-6-be/builds/1282/steps/vm%20tests/logs/stdio, for example.
@whesse , is this something you can help with or can give pointers where the the config files that should be fixed please? | infrastructure | two buildbots attempt to run test py with unrecognized failure summary somehow two sharded buildbots out of try to run tests with failure summary which fail because of that compare to for example whesse is this something you can help with or can give pointers where the the config files that should be fixed please | 1 |
1,028 | 3,023,875,747 | IssuesEvent | 2015-08-02 00:22:41 | catapult-project/catapult | https://api.github.com/repos/catapult-project/catapult | closed | Chromium check_licenses unhappy with third_party/tvcm files | Infrastructure P2 | <a href="https://github.com/natduca"><img src="https://avatars.githubusercontent.com/u/412396?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [natduca](https://github.com/natduca)**
_Monday Sep 22, 2014 at 20:23 GMT_
_Originally opened as https://github.com/google/trace-viewer/issues/528_
----
_From [aval...@chromium.org](https://code.google.com/u/103542635643954502142/) on February 03, 2014 08:13:24_
The check that broke/reverted the Chrome deps roll mentioned
third_party/trace-viewer/third_party/tvcm/tvcm/test_runner.py
having an unknown license.
_Original issue: http://code.google.com/p/trace-viewer/issues/detail?id=523_
| 1.0 | Chromium check_licenses unhappy with third_party/tvcm files - <a href="https://github.com/natduca"><img src="https://avatars.githubusercontent.com/u/412396?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [natduca](https://github.com/natduca)**
_Monday Sep 22, 2014 at 20:23 GMT_
_Originally opened as https://github.com/google/trace-viewer/issues/528_
----
_From [aval...@chromium.org](https://code.google.com/u/103542635643954502142/) on February 03, 2014 08:13:24_
The check that broke/reverted the Chrome deps roll mentioned
third_party/trace-viewer/third_party/tvcm/tvcm/test_runner.py
having an unknown license.
_Original issue: http://code.google.com/p/trace-viewer/issues/detail?id=523_
| infrastructure | chromium check licenses unhappy with third party tvcm files issue by monday sep at gmt originally opened as from on february the check that broke reverted the chrome deps roll mentioned third party trace viewer third party tvcm tvcm test runner py having an unknown license original issue | 1 |
103,141 | 12,865,902,786 | IssuesEvent | 2020-07-10 01:55:40 | lazy-suzy/lazysuzy-code | https://api.github.com/repos/lazy-suzy/lazysuzy-code | closed | Design Board | design prio: high | ### Board Updates
See updated Board moqup here: https://app.moqups.com/5NlUwWnQDj/edit/page/ac4ab0d91
- [x] 1. Extend Board to full width of the page with equal spacing on either side
- [x] 2. Use full height of window
- [x] 3. Update icon positioning as displayed
- [x] 4. Add 'Board Title' text input bar
- [x] 5. Update Board Catalog as per moqup here: https://app.moqups.com/5NlUwWnQDj/edit/page/af086e35d
- [x] 6. No scroll bars anywhere
- [x] 7. Move publish, share and account icons to the top right of page
- [x] 8. Fix Publish button font color
- [x] 9. Fix Search Bar styling: a) Rounded Corners b) square shape for Search icon background
- [x] 10. Add more space between search bar and tabs
- [x] 11. Clicking on "+" should open Browse Tab with My Items / Add New buttons
- [x] 12. My Items / Add New button shouldn't appear in Browse tab by default
- [x] 13. Delete confirmation window not yet visible
- [x] 14. Clicking on "+" icon on top right menu should open Board Config menu
- [x] 15. Open Filters window below Search Bar & Filters icon
- [x] 16. Is Filter window set up for tabs?
- [x] 17. Fix background colors (Board area should be #ffffff and surrounding area should be #f2f4f3
- [x] 18. Extend Tabs across width of side panel
- [x] 19. Remove box around "Board Title" when user input is not active
- [x] 20. Remove space between Icons and Board here and equalize spacing between Icons:
- [x] 21. Increase icon size and spacing here and extend the grey menu bar all the way to the end of the screen:
- [x] 22. Fix call out positioning so it doesn't run over the screen
- [x] 23. Increase and equalize spacing between Icons and Middle Align (currently looks Top Aligned)
- [x] 24. Center the text within the window.
- [x] 25. If switching back to Select or Browse from Board, raise the window back to the upper height
- [x] 26. Keep Board Title to the left of Icons and reduce spacing between Board icons on Mobile
- [x] 27. Fix alignment and spacing of objects on header menu. Expand grey bar for product options to full width of screen.
- [x] 28. Move Price to the right of Name/Brand and remove spacing between Product Name and Brand (for both Desktop and Mobile)
- [x] 29. Increase width of left panel nav bar
- [x] 30. **Select** panel looks nothing like the moqup.
- All category buttons should be the same size
- Category labels should not be on the same line as the category icon
- Filter window is appearing for some reason in the Select panel (see screenshot)
- Increase padding between category buttons on the same row
- [x] 31. **Browse** panel is missing Category title
- [x] 32. Product modal window should not move with scroll bar
- [x] 33. Update product modal window to match moqup. Position of _add_ and _like_ buttons should not move around.
- [x] 34. Mobile screenshot. Fix.
- [x] 35. Board tab
- Update title as per moqup
- Add more space between header and content
- Add more padding on left of item list in the panel
- [x] 36. Board title - Add input box around Board title on mouse hover
- [x] 37. Text icons
- Set equal spacing between icons
- Missing color for font color
- Conform label call-out spacing to match that for the image icons
- Case toggle icon looks smaller than others
- [x] 38. Update mobile menu as per moqup here: https://app.moqups.com/5NlUwWnQDj/edit/page/a3b3b1f58
Use > and < arrow to scroll between text menu options in a single row
- [x] 39. Make the _Board Name_ box on hover be transparent with border in #d6d6d6 and make sure the box does not touch the design board (see screenshot below)
- [x] 40. Fix styling for menu scroll arrow as per moqup to be white background, no border and #999999 color for arrow (see moqup here: https://app.moqups.com/5NlUwWnQDj/view/page/a5069bd4b)
- [x] 41. What is the grey box for next to the font size? Can we remove? Also scrolling for the menu bar should be one item at a time rather than the full row.
- [x] 42. Switch the top menu to scrolling menu on desktop if running out of space due to window resize rather than dropping in to multiple lines
- [x] 43. Update Select tab per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/aa0980e6c
- add header title (fixed)
- [x] 44. Update Browse tab per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/ac4ab0d91
- add back button next to Header title (return user to Select tab on click)
- add pre-defined filters (Sale, New, Best Sellers) - use _Click_ action to toggle between Selected and Unselected style
- Use "<" for Back button
- [x] 45. Add UI for price filter
- Change price label color and the price bar color to # b76e79
- [x] 46. Update open window icon in product modal window as per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/ac4ab0d91
- [x] 47. Add option to remove Favorites
- Nothing happens when I click on heart icon?
- [x] 48. Update text input per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/afb79b70c
- Increase font size and line spacing
- [x] 49. Update My Items page as per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/a1ace1e1d
- Add background to 'My Items' and 'All Uploads' options on mouse hover as per moqup
- Add instruction call-out as per screenshot below
- [x] 50. Update Screen Two and Three here: https://app.moqups.com/5NlUwWnQDj/edit/page/a954c051d
Screen Three:
- Increase spacing between rows
- Left align "Keep private" check box
- [x] 51. Clicking on 'Add via URL' button should directly open to the below window
- [x] 52. Fix scroll bar issue in content bar
- [x] 53. Reduce spacing here

- [x] 54. On Favorites page:
- remove heart icon from images
- add product detail bar at bottom of content menu
- show heart colored in for products on here in detail bar
[UPDATE]: clicking on any item here is not opening up the product detail window at the bottom
- [x] 55. Only product listing section should scroll in content menu
[UPDATE]:
- How do I test this on desktop view? Why is the page only pulling 15 products and not more?
Mobile view comments (see screenshot):
- Product detail window is not appearing properly
- Scroll-bar should appear in product window section only
- Add more spacing between the filters and product listing section
- Overall, look at how neat the layout appears in the second screenshot below vs. ours looks very raw as if we just cobbled bunch of elements together


- [x] 56. Change call out for gear icon from "New Board" to "Edit Board settings"
- [x] 57. Header of content section for _Browse_ tab should match the Category selected. Currently is hard coded to "Sofa" no matter what Category is chosen.
**Add Image Menu**
- [x] 1. Position Upload Image / Add Image by URL buttons in middle of window
- [x] 2. Position Image URL input in middle of window
- [x] 3. Fix text in "Must be .png or .jpg"
- [x] 4. Remove curves at endpoints of input line
- [ ] 5. Remove input text on mouse click
- [x] 6. Move instruction text closer to input line
- [x] 7. Make image URL a required field
- [x] 8. Space out Title / Price / Tags input lines across the window
- [x] 9. Use large x for window close with thinner lines and fix color
- [x] 10. Add more space between Buttons and Step counter
- [x] 11. Close window on Add Button click
- [x] 12. Window appears cut-off on mobile. Is it being opened as modal window on mobile as well?

----------------------------
- [x] 1. Create Board Configuration menu
(M) https://app.moqups.com/5NlUwWnQDj/edit/page/a14bef619
For Desktop, create same menu in a pop-up window
- [x] 2. Clicking on **"+" Button** in Browse window should take user to the following screen:
(D) https://app.moqups.com/5NlUwWnQDj/edit/page/a2fa4a11f
(M) https://app.moqups.com/5NlUwWnQDj/edit/page/a5e429983
Clicking on **"Add New" Button** in the above screens should then take user to the add product menu
- [x] 3. Add Catalog Panel
(D) https://app.moqups.com/5NlUwWnQDj/edit/page/aa482ba12
(M) https://app.moqups.com/5NlUwWnQDj/edit/page/a5e429983
- [x] 4. Product Panel is missing Delete Button and Forward / Backward buttons
- [x] 5. Add Delete Confirmation pop-up when clicking on Delete Button in Canvas Panel
Confirmation window: https://app.moqups.com/5NlUwWnQDj/edit/page/a304f8c44
- [x] 6. Desktop: Move Canvas Panel to upper right corner of panel as per moqup and space out icons further | 1.0 | Design Board - ### Board Updates
See updated Board moqup here: https://app.moqups.com/5NlUwWnQDj/edit/page/ac4ab0d91
- [x] 1. Extend Board to full width of the page with equal spacing on either side
- [x] 2. Use full height of window
- [x] 3. Update icon positioning as displayed
- [x] 4. Add 'Board Title' text input bar
- [x] 5. Update Board Catalog as per moqup here: https://app.moqups.com/5NlUwWnQDj/edit/page/af086e35d
- [x] 6. No scroll bars anywhere
- [x] 7. Move publish, share and account icons to the top right of page
- [x] 8. Fix Publish button font color
- [x] 9. Fix Search Bar styling: a) Rounded Corners b) square shape for Search icon background
- [x] 10. Add more space between search bar and tabs
- [x] 11. Clicking on "+" should open Browse Tab with My Items / Add New buttons
- [x] 12. My Items / Add New button shouldn't appear in Browse tab by default
- [x] 13. Delete confirmation window not yet visible
- [x] 14. Clicking on "+" icon on top right menu should open Board Config menu
- [x] 15. Open Filters window below Search Bar & Filters icon
- [x] 16. Is Filter window set up for tabs?
- [x] 17. Fix background colors (Board area should be #ffffff and surrounding area should be #f2f4f3
- [x] 18. Extend Tabs across width of side panel
- [x] 19. Remove box around "Board Title" when user input is not active
- [x] 20. Remove space between Icons and Board here and equalize spacing between Icons:
- [x] 21. Increase icon size and spacing here and extend the grey menu bar all the way to the end of the screen:
- [x] 22. Fix call out positioning so it doesn't run over the screen
- [x] 23. Increase and equalize spacing between Icons and Middle Align (currently looks Top Aligned)
- [x] 24. Center the text within the window.
- [x] 25. If switching back to Select or Browse from Board, raise the window back to the upper height
- [x] 26. Keep Board Title to the left of Icons and reduce spacing between Board icons on Mobile
- [x] 27. Fix alignment and spacing of objects on header menu. Expand grey bar for product options to full width of screen.
- [x] 28. Move Price to the right of Name/Brand and remove spacing between Product Name and Brand (for both Desktop and Mobile)
- [x] 29. Increase width of left panel nav bar
- [x] 30. **Select** panel looks nothing like the moqup.
- All category buttons should be the same size
- Category labels should not be on the same line as the category icon
- Filter window is appearing for some reason in the Select panel (see screenshot)
- Increase padding between category buttons on the same row
- [x] 31. **Browse** panel is missing Category title
- [x] 32. Product modal window should not move with scroll bar
- [x] 33. Update product modal window to match moqup. Position of _add_ and _like_ buttons should not move around.
- [x] 34. Mobile screenshot. Fix.
- [x] 35. Board tab
- Update title as per moqup
- Add more space between header and content
- Add more padding on left of item list in the panel
- [x] 36. Board title - Add input box around Board title on mouse hover
- [x] 37. Text icons
- Set equal spacing between icons
- Missing color for font color
- Conform label call-out spacing to match that for the image icons
- Case toggle icon looks smaller than others
- [x] 38. Update mobile menu as per moqup here: https://app.moqups.com/5NlUwWnQDj/edit/page/a3b3b1f58
Use > and < arrow to scroll between text menu options in a single row
- [x] 39. Make the _Board Name_ box on hover be transparent with border in #d6d6d6 and make sure the box does not touch the design board (see screenshot below)
- [x] 40. Fix styling for menu scroll arrow as per moqup to be white background, no border and #999999 color for arrow (see moqup here: https://app.moqups.com/5NlUwWnQDj/view/page/a5069bd4b)
- [x] 41. What is the grey box for next to the font size? Can we remove? Also scrolling for the menu bar should be one item at a time rather than the full row.
- [x] 42. Switch the top menu to scrolling menu on desktop if running out of space due to window resize rather than dropping in to multiple lines
- [x] 43. Update Select tab per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/aa0980e6c
- add header title (fixed)
- [x] 44. Update Browse tab per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/ac4ab0d91
- add back button next to Header title (return user to Select tab on click)
- add pre-defined filters (Sale, New, Best Sellers) - use _Click_ action to toggle between Selected and Unselected style
- Use "<" for Back button
- [x] 45. Add UI for price filter
- Change price label color and the price bar color to # b76e79
- [x] 46. Update open window icon in product modal window as per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/ac4ab0d91
- [x] 47. Add option to remove Favorites
- Nothing happens when I click on heart icon?
- [x] 48. Update text input per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/afb79b70c
- Increase font size and line spacing
- [x] 49. Update My Items page as per moqup: https://app.moqups.com/5NlUwWnQDj/edit/page/a1ace1e1d
- Add background to 'My Items' and 'All Uploads' options on mouse hover as per moqup
- Add instruction call-out as per screenshot below
- [x] 50. Update Screen Two and Three here: https://app.moqups.com/5NlUwWnQDj/edit/page/a954c051d
Screen Three:
- Increase spacing between rows
- Left align "Keep private" check box
- [x] 51. Clicking on 'Add via URL' button should directly open to the below window
- [x] 52. Fix scroll bar issue in content bar
- [x] 53. Reduce spacing here

- [x] 54. On Favorites page:
- remove heart icon from images
- add product detail bar at bottom of content menu
- show heart colored in for products on here in detail bar
[UPDATE]: clicking on any item here is not opening up the product detail window at the bottom
- [x] 55. Only product listing section should scroll in content menu
[UPDATE]:
- How do I test this on desktop view? Why is the page only pulling 15 products and not more?
Mobile view comments (see screenshot):
- Product detail window is not appearing properly
- Scroll-bar should appear in product window section only
- Add more spacing between the filters and product listing section
- Overall, look at how neat the layout appears in the second screenshot below vs. ours looks very raw as if we just cobbled bunch of elements together


- [x] 56. Change call out for gear icon from "New Board" to "Edit Board settings"
- [x] 57. Header of content section for _Browse_ tab should match the Category selected. Currently is hard coded to "Sofa" no matter what Category is chosen.
**Add Image Menu**
- [x] 1. Position Upload Image / Add Image by URL buttons in middle of window
- [x] 2. Position Image URL input in middle of window
- [x] 3. Fix text in "Must be .png or .jpg"
- [x] 4. Remove curves at endpoints of input line
- [ ] 5. Remove input text on mouse click
- [x] 6. Move instruction text closer to input line
- [x] 7. Make image URL a required field
- [x] 8. Space out Title / Price / Tags input lines across the window
- [x] 9. Use large x for window close with thinner lines and fix color
- [x] 10. Add more space between Buttons and Step counter
- [x] 11. Close window on Add Button click
- [x] 12. Window appears cut-off on mobile. Is it being opened as modal window on mobile as well?

----------------------------
- [x] 1. Create Board Configuration menu
(M) https://app.moqups.com/5NlUwWnQDj/edit/page/a14bef619
For Desktop, create same menu in a pop-up window
- [x] 2. Clicking on **"+" Button** in Browse window should take user to the following screen:
(D) https://app.moqups.com/5NlUwWnQDj/edit/page/a2fa4a11f
(M) https://app.moqups.com/5NlUwWnQDj/edit/page/a5e429983
Clicking on **"Add New" Button** in the above screens should then take user to the add product menu
- [x] 3. Add Catalog Panel
(D) https://app.moqups.com/5NlUwWnQDj/edit/page/aa482ba12
(M) https://app.moqups.com/5NlUwWnQDj/edit/page/a5e429983
- [x] 4. Product Panel is missing Delete Button and Forward / Backward buttons
- [x] 5. Add Delete Confirmation pop-up when clicking on Delete Button in Canvas Panel
Confirmation window: https://app.moqups.com/5NlUwWnQDj/edit/page/a304f8c44
- [x] 6. Desktop: Move Canvas Panel to upper right corner of panel as per moqup and space out icons further | non_infrastructure | design board board updates see updated board moqup here extend board to full width of the page with equal spacing on either side use full height of window update icon positioning as displayed add board title text input bar update board catalog as per moqup here no scroll bars anywhere move publish share and account icons to the top right of page fix publish button font color fix search bar styling a rounded corners b square shape for search icon background add more space between search bar and tabs clicking on should open browse tab with my items add new buttons my items add new button shouldn t appear in browse tab by default delete confirmation window not yet visible clicking on icon on top right menu should open board config menu open filters window below search bar filters icon is filter window set up for tabs fix background colors board area should be ffffff and surrounding area should be extend tabs across width of side panel remove box around board title when user input is not active remove space between icons and board here and equalize spacing between icons increase icon size and spacing here and extend the grey menu bar all the way to the end of the screen fix call out positioning so it doesn t run over the screen increase and equalize spacing between icons and middle align currently looks top aligned center the text within the window if switching back to select or browse from board raise the window back to the upper height keep board title to the left of icons and reduce spacing between board icons on mobile fix alignment and spacing of objects on header menu expand grey bar for product options to full width of screen move price to the right of name brand and remove spacing between product name and brand for both desktop and mobile increase width of left panel nav bar select panel looks nothing like the moqup all category buttons should be the same size category labels should not be on the same line as the category icon filter window is appearing for some reason in the select panel see screenshot increase padding between category buttons on the same row browse panel is missing category title product modal window should not move with scroll bar update product modal window to match moqup position of add and like buttons should not move around mobile screenshot fix board tab update title as per moqup add more space between header and content add more padding on left of item list in the panel board title add input box around board title on mouse hover text icons set equal spacing between icons missing color for font color conform label call out spacing to match that for the image icons case toggle icon looks smaller than others update mobile menu as per moqup here use and arrow to scroll between text menu options in a single row make the board name box on hover be transparent with border in and make sure the box does not touch the design board see screenshot below fix styling for menu scroll arrow as per moqup to be white background no border and color for arrow see moqup here what is the grey box for next to the font size can we remove also scrolling for the menu bar should be one item at a time rather than the full row switch the top menu to scrolling menu on desktop if running out of space due to window resize rather than dropping in to multiple lines update select tab per moqup add header title fixed update browse tab per moqup add back button next to header title return user to select tab on click add pre defined filters sale new best sellers use click action to toggle between selected and unselected style use for back button add ui for price filter change price label color and the price bar color to update open window icon in product modal window as per moqup add option to remove favorites nothing happens when i click on heart icon update text input per moqup increase font size and line spacing update my items page as per moqup add background to my items and all uploads options on mouse hover as per moqup add instruction call out as per screenshot below update screen two and three here screen three increase spacing between rows left align keep private check box clicking on add via url button should directly open to the below window fix scroll bar issue in content bar reduce spacing here on favorites page remove heart icon from images add product detail bar at bottom of content menu show heart colored in for products on here in detail bar clicking on any item here is not opening up the product detail window at the bottom only product listing section should scroll in content menu how do i test this on desktop view why is the page only pulling products and not more mobile view comments see screenshot product detail window is not appearing properly scroll bar should appear in product window section only add more spacing between the filters and product listing section overall look at how neat the layout appears in the second screenshot below vs ours looks very raw as if we just cobbled bunch of elements together change call out for gear icon from new board to edit board settings header of content section for browse tab should match the category selected currently is hard coded to sofa no matter what category is chosen add image menu position upload image add image by url buttons in middle of window position image url input in middle of window fix text in must be png or jpg remove curves at endpoints of input line remove input text on mouse click move instruction text closer to input line make image url a required field space out title price tags input lines across the window use large x for window close with thinner lines and fix color add more space between buttons and step counter close window on add button click window appears cut off on mobile is it being opened as modal window on mobile as well create board configuration menu m for desktop create same menu in a pop up window clicking on button in browse window should take user to the following screen d m clicking on add new button in the above screens should then take user to the add product menu add catalog panel d m product panel is missing delete button and forward backward buttons add delete confirmation pop up when clicking on delete button in canvas panel confirmation window desktop move canvas panel to upper right corner of panel as per moqup and space out icons further | 0 |
20,683 | 14,099,402,167 | IssuesEvent | 2020-11-06 01:17:47 | noahtalerman/test-issues-kolide | https://api.github.com/repos/noahtalerman/test-issues-kolide | opened | [CLOSED] Turning the GC off at build-time on Windows | Component: Development Infrastructure | <a href="https://github.com/marpaia"><img src="https://avatars2.githubusercontent.com/u/927168?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [marpaia](https://github.com/marpaia)**
_Wednesday Oct 12, 2016 at 16:19 GMT_
_Originally opened as https://github.com/kolide/fleet/pull/303_
----
The way that you define environment variables for a single command
is slightly different than how it's done on nix. This PR adds a simple
switch to the Makefile to use the correct approach on the appropriate
operating system.
----
_**[marpaia](https://github.com/marpaia)** included the following code: https://github.com/kolide/fleet/pull/303/commits_
| 1.0 | [CLOSED] Turning the GC off at build-time on Windows - <a href="https://github.com/marpaia"><img src="https://avatars2.githubusercontent.com/u/927168?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [marpaia](https://github.com/marpaia)**
_Wednesday Oct 12, 2016 at 16:19 GMT_
_Originally opened as https://github.com/kolide/fleet/pull/303_
----
The way that you define environment variables for a single command
is slightly different than how it's done on nix. This PR adds a simple
switch to the Makefile to use the correct approach on the appropriate
operating system.
----
_**[marpaia](https://github.com/marpaia)** included the following code: https://github.com/kolide/fleet/pull/303/commits_
| infrastructure | turning the gc off at build time on windows issue by wednesday oct at gmt originally opened as the way that you define environment variables for a single command is slightly different than how it s done on nix this pr adds a simple switch to the makefile to use the correct approach on the appropriate operating system included the following code | 1 |
9,141 | 7,840,870,960 | IssuesEvent | 2018-06-18 17:45:28 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Building from source fails to download required dotnet cli tool with 404 not found error. | area-Infrastructure bug | Trying to build CoreFX from source by running build scripts fails to download dotnet cli tool with 404 Not Found.
Here is error on cmd:
```
Installing dotnet cli...
ERROR: Could not install dotnet cli correctly.
Please check the detailed log that follows.
Running E:\GitHub\corefx\init-tools.cmd
Installing 'https://dotnetcli.azureedge.net/dotnet/Sdk/2.1.300-rc1-008673/dotnet-sdk-2.1.300-rc1-008673-win-x64.zip' to 'E:\GitHub\corefx\Tools\dotnetcli\dotnet-sdk-2.1.300-rc1-008673-win-x64.zip'
Exception calling "DownloadFile" with "2" argument(s): "The remote server returned an error: (404) Not Found."
At line:1 char:249
+ ... dentials; } $wc.DownloadFile('https://dotnetcli.azureedge.net/dotnet/ ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : NotSpecified: (:) [], ParentContainsErrorRecordException
+ FullyQualifiedErrorId : WebException
```
Also. I tried to access download URL from browser and fails with 404 Not Found, Here is output from browser:
````
<Error>
<Code>ResourceNotFound</Code>
<Message>
The specified resource does not exist. RequestId:3aca3efb-b01e-003a-4e9b-06eb68000000 Time:2018-06-18T00:32:14.4887599Z
</Message>
</Error>
````
I have direct access to the internet (no proxy).
**Location details:**
region: North Africa
country: Algeria
cc @karelz | 1.0 | Building from source fails to download required dotnet cli tool with 404 not found error. - Trying to build CoreFX from source by running build scripts fails to download dotnet cli tool with 404 Not Found.
Here is error on cmd:
```
Installing dotnet cli...
ERROR: Could not install dotnet cli correctly.
Please check the detailed log that follows.
Running E:\GitHub\corefx\init-tools.cmd
Installing 'https://dotnetcli.azureedge.net/dotnet/Sdk/2.1.300-rc1-008673/dotnet-sdk-2.1.300-rc1-008673-win-x64.zip' to 'E:\GitHub\corefx\Tools\dotnetcli\dotnet-sdk-2.1.300-rc1-008673-win-x64.zip'
Exception calling "DownloadFile" with "2" argument(s): "The remote server returned an error: (404) Not Found."
At line:1 char:249
+ ... dentials; } $wc.DownloadFile('https://dotnetcli.azureedge.net/dotnet/ ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : NotSpecified: (:) [], ParentContainsErrorRecordException
+ FullyQualifiedErrorId : WebException
```
Also. I tried to access download URL from browser and fails with 404 Not Found, Here is output from browser:
````
<Error>
<Code>ResourceNotFound</Code>
<Message>
The specified resource does not exist. RequestId:3aca3efb-b01e-003a-4e9b-06eb68000000 Time:2018-06-18T00:32:14.4887599Z
</Message>
</Error>
````
I have direct access to the internet (no proxy).
**Location details:**
region: North Africa
country: Algeria
cc @karelz | infrastructure | building from source fails to download required dotnet cli tool with not found error trying to build corefx from source by running build scripts fails to download dotnet cli tool with not found here is error on cmd installing dotnet cli error could not install dotnet cli correctly please check the detailed log that follows running e github corefx init tools cmd installing to e github corefx tools dotnetcli dotnet sdk win zip exception calling downloadfile with argument s the remote server returned an error not found at line char dentials wc downloadfile categoryinfo notspecified parentcontainserrorrecordexception fullyqualifiederrorid webexception also i tried to access download url from browser and fails with not found here is output from browser resourcenotfound the specified resource does not exist requestid time i have direct access to the internet no proxy location details region north africa country algeria cc karelz | 1 |
940 | 3,006,286,101 | IssuesEvent | 2015-07-27 09:24:35 | Itseez/opencv | https://api.github.com/repos/Itseez/opencv | opened | Clean up test data | auto-transferred category: infrastructure feature priority: normal | Transferred from http://code.opencv.org/issues/2722
```
|| Kirill Kornyakov on 2013-01-21 13:43
|| Priority: Normal
|| Affected: None
|| Category: infrastructure
|| Tracker: Feature
|| Difficulty: None
|| PR:
|| Platform: None / None
```
Clean up test data
-----------
```
On a weekly OpenCV meeting we decided, that we need to clean & minimize test data. Here are tasks selected for the hackathon:
# Remove unused images and duplicated files (cascades)
# Look at large files and decide if we need the at all
# Make tests to use the same input images, i.e. stitching images can be used by feature2d tests
# Remove auto-generated images, but if only their size is large
```
History
-------
##### Vadim Pisarevsky on 2013-01-24 17:54
```
- Assignee deleted (Vadim Pisarevsky)
```
##### Kirill Kornyakov on 2013-02-14 08:08
```
- Target version set to 2.4.5
```
##### Andrey Kamaev on 2013-04-01 09:56
```
- Target version changed from 2.4.5 to 3.0
``` | 1.0 | Clean up test data - Transferred from http://code.opencv.org/issues/2722
```
|| Kirill Kornyakov on 2013-01-21 13:43
|| Priority: Normal
|| Affected: None
|| Category: infrastructure
|| Tracker: Feature
|| Difficulty: None
|| PR:
|| Platform: None / None
```
Clean up test data
-----------
```
On a weekly OpenCV meeting we decided, that we need to clean & minimize test data. Here are tasks selected for the hackathon:
# Remove unused images and duplicated files (cascades)
# Look at large files and decide if we need the at all
# Make tests to use the same input images, i.e. stitching images can be used by feature2d tests
# Remove auto-generated images, but if only their size is large
```
History
-------
##### Vadim Pisarevsky on 2013-01-24 17:54
```
- Assignee deleted (Vadim Pisarevsky)
```
##### Kirill Kornyakov on 2013-02-14 08:08
```
- Target version set to 2.4.5
```
##### Andrey Kamaev on 2013-04-01 09:56
```
- Target version changed from 2.4.5 to 3.0
``` | infrastructure | clean up test data transferred from kirill kornyakov on priority normal affected none category infrastructure tracker feature difficulty none pr platform none none clean up test data on a weekly opencv meeting we decided that we need to clean minimize test data here are tasks selected for the hackathon remove unused images and duplicated files cascades look at large files and decide if we need the at all make tests to use the same input images i e stitching images can be used by tests remove auto generated images but if only their size is large history vadim pisarevsky on assignee deleted vadim pisarevsky kirill kornyakov on target version set to andrey kamaev on target version changed from to | 1 |
27,956 | 12,741,276,346 | IssuesEvent | 2020-06-26 05:35:00 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | reopened | Learn about Public Works' project tracking workflows | Epic Product: Mobility Project Database Service: Product Type: Research Workgroup: ATD Workgroup: PWD | [Urban Trails Program](https://www.austintexas.gov/urbantrails)
- Staff contact: Katie W.
[Safe Routes to School](http://www.austintexas.gov/department/child-safety-program)
- Staff contact: Aleksiina
- [Notes](https://drive.google.com/open?id=1DfEwo3z3rwNXwTF2XpbRqzulTBt9gIbC7ML_MJmVDgM)
[Sidewalks](http://www.austintexas.gov/sidewalks)
- Staff contact:
*Migrated from [atd-mobility-project-database #40](https://github.com/cityofaustin/atd-mobility-project-database/issues/40)* | 1.0 | Learn about Public Works' project tracking workflows - [Urban Trails Program](https://www.austintexas.gov/urbantrails)
- Staff contact: Katie W.
[Safe Routes to School](http://www.austintexas.gov/department/child-safety-program)
- Staff contact: Aleksiina
- [Notes](https://drive.google.com/open?id=1DfEwo3z3rwNXwTF2XpbRqzulTBt9gIbC7ML_MJmVDgM)
[Sidewalks](http://www.austintexas.gov/sidewalks)
- Staff contact:
*Migrated from [atd-mobility-project-database #40](https://github.com/cityofaustin/atd-mobility-project-database/issues/40)* | non_infrastructure | learn about public works project tracking workflows staff contact katie w staff contact aleksiina staff contact migrated from | 0 |
93,679 | 27,014,045,473 | IssuesEvent | 2023-02-10 17:40:25 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | Crossgen2 asserts in JIT when compiling The JIT\Regression\JitBlue\Runtime_81739 | blocking-clean-ci Known Build Error | ## Build Information
Build: https://dev.azure.com/dnceng-public/cbb18261-c48f-4abb-8651-8cdcb5474649/_build/results?buildId=167278
Build error leg or test failing: JIT\\Regression\\JitBlue\\Runtime_81739\\Runtime_81739\\Runtime_81739.cmd
Pull request: https://github.com/dotnet/runtime/pull/81956
<!-- Error message template -->
## Error Message
" "dotnet" "C:\h\w\ACF5090A\p\crossgen2\crossgen2.dll" @"C:\h\w\ACF5090A\w\AA3A0904\e\JIT\Regression\JitBlue\Runtime_81739\Runtime_81739\Runtime_81739.dll.rsp" -r:C:\h\w\ACF5090A\w\AA3A0904\e\JIT\Regression\JitBlue\Runtime_81739\Runtime_81739\IL-CG2\*.dll"
D:\a\_work\1\s\src\coreclr\jit\emitxarch.cpp:6034
Assertion failed 'isFloatReg(dstReg) && isFloatReg(srcReg)' in 'Runtime_81739:Main():int' during 'Generate code' (IL size 27; hash 0x0dd256ab; FullOpts)
| 1.0 | Crossgen2 asserts in JIT when compiling The JIT\Regression\JitBlue\Runtime_81739 - ## Build Information
Build: https://dev.azure.com/dnceng-public/cbb18261-c48f-4abb-8651-8cdcb5474649/_build/results?buildId=167278
Build error leg or test failing: JIT\\Regression\\JitBlue\\Runtime_81739\\Runtime_81739\\Runtime_81739.cmd
Pull request: https://github.com/dotnet/runtime/pull/81956
<!-- Error message template -->
## Error Message
" "dotnet" "C:\h\w\ACF5090A\p\crossgen2\crossgen2.dll" @"C:\h\w\ACF5090A\w\AA3A0904\e\JIT\Regression\JitBlue\Runtime_81739\Runtime_81739\Runtime_81739.dll.rsp" -r:C:\h\w\ACF5090A\w\AA3A0904\e\JIT\Regression\JitBlue\Runtime_81739\Runtime_81739\IL-CG2\*.dll"
D:\a\_work\1\s\src\coreclr\jit\emitxarch.cpp:6034
Assertion failed 'isFloatReg(dstReg) && isFloatReg(srcReg)' in 'Runtime_81739:Main():int' during 'Generate code' (IL size 27; hash 0x0dd256ab; FullOpts)
| non_infrastructure | asserts in jit when compiling the jit regression jitblue runtime build information build build error leg or test failing jit regression jitblue runtime runtime runtime cmd pull request error message dotnet c h w p dll c h w w e jit regression jitblue runtime runtime runtime dll rsp r c h w w e jit regression jitblue runtime runtime il dll d a work s src coreclr jit emitxarch cpp assertion failed isfloatreg dstreg isfloatreg srcreg in runtime main int during generate code il size hash fullopts | 0 |
134,215 | 5,222,227,026 | IssuesEvent | 2017-01-27 07:00:25 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Articles of food (and potentially pets) are not translatable | priority: medium status: issue: in progress type: translations | I'm creating a new issue in addition to mentioning this bug in https://github.com/HabitRPG/habitrpg/issues/4522#issuecomment-110843828 because it seems not so trivial at first glance:
articles ("a") of food and pets' names are not translatable in [content.coffee](https://github.com/HabitRPG/habitrpg/blob/develop/common/script/content.coffee#L1080) thus [pets.json feedPet string](https://github.com/HabitRPG/habitrpg/blob/develop/common/locales/en/pets.json#L54) (`Feed <%= article %><%= text %> to your <%= name %>?`) rendered in [inventoryCtrl.js](https://github.com/HabitRPG/habitrpg/blob/develop/website/public/js/controllers/inventoryCtrl.js#L126) returns a mix of translated text and untranslatable article "a" in the middle
Initially reported by @ルナ in the guild chat
Slightly connected https://github.com/HabitRPG/habitrpg/issues/3571
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/24517631-articles-of-food-and-potentially-pets-are-not-translatable?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
| 1.0 | Articles of food (and potentially pets) are not translatable - I'm creating a new issue in addition to mentioning this bug in https://github.com/HabitRPG/habitrpg/issues/4522#issuecomment-110843828 because it seems not so trivial at first glance:
articles ("a") of food and pets' names are not translatable in [content.coffee](https://github.com/HabitRPG/habitrpg/blob/develop/common/script/content.coffee#L1080) thus [pets.json feedPet string](https://github.com/HabitRPG/habitrpg/blob/develop/common/locales/en/pets.json#L54) (`Feed <%= article %><%= text %> to your <%= name %>?`) rendered in [inventoryCtrl.js](https://github.com/HabitRPG/habitrpg/blob/develop/website/public/js/controllers/inventoryCtrl.js#L126) returns a mix of translated text and untranslatable article "a" in the middle
Initially reported by @ルナ in the guild chat
Slightly connected https://github.com/HabitRPG/habitrpg/issues/3571
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/24517631-articles-of-food-and-potentially-pets-are-not-translatable?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
| non_infrastructure | articles of food and potentially pets are not translatable i m creating a new issue in addition to mentioning this bug in because it seems not so trivial at first glance articles a of food and pets names are not translatable in thus feed to your rendered in returns a mix of translated text and untranslatable article a in the middle initially reported by ルナ in the guild chat slightly connected want to back this issue we accept bounties via | 0 |
13,559 | 10,322,141,897 | IssuesEvent | 2019-08-31 09:21:21 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | NuGet package `microsoft.netcore.app` version 3.0+ removes reference assemblies | area-Infrastructure question | In 2.x versions we have, for example, `ref\netcoreapp2.2\netstandard.dll` -- that's easy to understand, and we can just feed the assemblies in this folder to the compiler services for type checking etc.
But now in 3.0 preview packages this folder is gone. The compiler service complains about missing types.
Where should we look now? | 1.0 | NuGet package `microsoft.netcore.app` version 3.0+ removes reference assemblies - In 2.x versions we have, for example, `ref\netcoreapp2.2\netstandard.dll` -- that's easy to understand, and we can just feed the assemblies in this folder to the compiler services for type checking etc.
But now in 3.0 preview packages this folder is gone. The compiler service complains about missing types.
Where should we look now? | infrastructure | nuget package microsoft netcore app version removes reference assemblies in x versions we have for example ref netstandard dll that s easy to understand and we can just feed the assemblies in this folder to the compiler services for type checking etc but now in preview packages this folder is gone the compiler service complains about missing types where should we look now | 1 |
31,598 | 25,919,561,111 | IssuesEvent | 2022-12-15 20:31:12 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | closed | Update bundled WP for tests | P1 QA: Eng Type: Infrastructure | ## Feature Description
<!-- Please describe clear and concisely which problem the feature would solve or which publisher needs it would address. -->
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* The bundled version of WordPress used for tests locally should be updated to the latest version, including test-related dependencies
## Implementation Brief
* Increase Composer's `platform.php` config to `5.6.20` as required by `roots/wordpress-no-content` (`roots/wordpress` is a metapackage that in turn requires this) – not sure if something in Composer changed regarding this or what because it won't select the package with the current version constraints due to our version using `5.6` which it sees as `5.6.0` rather than `5.6.*` which is not how the platform config works. See https://getcomposer.org/doc/06-config.md#platform
* Run `composer update roots/wordpress wp-phpunit/wp-phpunit --with-all-dependencies` (the flag only updates dependencies of the listed packages)
### Test Coverage
* No changes, this only affects local development
* Ensure `composer test` runs successfully locally
## QA Brief
* This change only really affects tests locally but could potentially cause a different package to be selected so the changes to the lock file should be reviewed as well as ensuring tests run locally via `composer test` (works for me!)
## Changelog entry
* N/A
| 1.0 | Update bundled WP for tests - ## Feature Description
<!-- Please describe clear and concisely which problem the feature would solve or which publisher needs it would address. -->
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* The bundled version of WordPress used for tests locally should be updated to the latest version, including test-related dependencies
## Implementation Brief
* Increase Composer's `platform.php` config to `5.6.20` as required by `roots/wordpress-no-content` (`roots/wordpress` is a metapackage that in turn requires this) – not sure if something in Composer changed regarding this or what because it won't select the package with the current version constraints due to our version using `5.6` which it sees as `5.6.0` rather than `5.6.*` which is not how the platform config works. See https://getcomposer.org/doc/06-config.md#platform
* Run `composer update roots/wordpress wp-phpunit/wp-phpunit --with-all-dependencies` (the flag only updates dependencies of the listed packages)
### Test Coverage
* No changes, this only affects local development
* Ensure `composer test` runs successfully locally
## QA Brief
* This change only really affects tests locally but could potentially cause a different package to be selected so the changes to the lock file should be reviewed as well as ensuring tests run locally via `composer test` (works for me!)
## Changelog entry
* N/A
| infrastructure | update bundled wp for tests feature description do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria the bundled version of wordpress used for tests locally should be updated to the latest version including test related dependencies implementation brief increase composer s platform php config to as required by roots wordpress no content roots wordpress is a metapackage that in turn requires this – not sure if something in composer changed regarding this or what because it won t select the package with the current version constraints due to our version using which it sees as rather than which is not how the platform config works see run composer update roots wordpress wp phpunit wp phpunit with all dependencies the flag only updates dependencies of the listed packages test coverage no changes this only affects local development ensure composer test runs successfully locally qa brief this change only really affects tests locally but could potentially cause a different package to be selected so the changes to the lock file should be reviewed as well as ensuring tests run locally via composer test works for me changelog entry n a | 1 |
32,558 | 26,783,722,023 | IssuesEvent | 2023-02-01 00:00:06 | dart-lang/site-www | https://api.github.com/repos/dart-lang/site-www | closed | Verify site and all code snippets are prepared to build/run with Dart 3 | infrastructure p1-high e1-hours e2-days | ### Describe the problem
We should make sure the site and all examples build and run with Dart 3 (with sound null safety) and make any updates as necessary.
### Expected fix
- [ ] Build, run, analyze, and test all site scripts and examples
- [ ] Make any necessary updates
- [x] Migrate `package:linkcheck` to run with sound null safety (https://github.com/filiph/linkcheck/pull/105)
### Additional context
_No response_ | 1.0 | Verify site and all code snippets are prepared to build/run with Dart 3 - ### Describe the problem
We should make sure the site and all examples build and run with Dart 3 (with sound null safety) and make any updates as necessary.
### Expected fix
- [ ] Build, run, analyze, and test all site scripts and examples
- [ ] Make any necessary updates
- [x] Migrate `package:linkcheck` to run with sound null safety (https://github.com/filiph/linkcheck/pull/105)
### Additional context
_No response_ | infrastructure | verify site and all code snippets are prepared to build run with dart describe the problem we should make sure the site and all examples build and run with dart with sound null safety and make any updates as necessary expected fix build run analyze and test all site scripts and examples make any necessary updates migrate package linkcheck to run with sound null safety additional context no response | 1 |
652,729 | 21,559,882,145 | IssuesEvent | 2022-05-01 02:11:47 | MaksiRose/paw-and-paper | https://api.github.com/repos/MaksiRose/paw-and-paper | closed | Bug using feature friendship: First heart doesn't send message | bug PRIORITY | ### Current behavior
When you get your first heart, no message is sent.
### Expected behavior
A message should be sent saying that the friendship grew
### Steps to reproduce the behavior
1. Ping someone for the first time
2. Check rp friendships
3. Get to 3 friendship points
4. Now it will say that it grew to 2 hearts
### Additional details
_No response_ | 1.0 | Bug using feature friendship: First heart doesn't send message - ### Current behavior
When you get your first heart, no message is sent.
### Expected behavior
A message should be sent saying that the friendship grew
### Steps to reproduce the behavior
1. Ping someone for the first time
2. Check rp friendships
3. Get to 3 friendship points
4. Now it will say that it grew to 2 hearts
### Additional details
_No response_ | non_infrastructure | bug using feature friendship first heart doesn t send message current behavior when you get your first heart no message is sent expected behavior a message should be sent saying that the friendship grew steps to reproduce the behavior ping someone for the first time check rp friendships get to friendship points now it will say that it grew to hearts additional details no response | 0 |
546 | 2,502,395,506 | IssuesEvent | 2015-01-09 08:26:07 | fossology/fossology | https://api.github.com/repos/fossology/fossology | opened | wget_agent uses -v switch to tar (twice) and then throws all the output away | Component: Rank Component: Tester Priority: Normal Status: New Tracker: Bug | ---
Author Name: **Gervase Markham**
Original Redmine Issue: 7335, http://www.fossology.org/issues/7335
Original Date: 2014/07/23
---
gerv@hare:/usr/src/fossology$ ack "cvvf"
src/wget_agent/agent/wget_agent.c
404: snprintf(CMD,MAXCMD-1, "tar -cvvf '%s' -C '%s' ./ >/dev/null 2>&1", TempFile, TempFilePath);
505: snprintf(command,MAXCMD-1, "tar -cvvf '%s' -C '%s' ./ >/dev/null 2>&1", GlobalTempFile, TempFileDirectory);
669: snprintf(CMD,MAXCMD-1, "tar -cvvf '%s' -C '%s' ./ %s >/dev/null 2>&1", TempFile, Path, GlobalParam);
689: snprintf(CMD,MAXCMD-1, "tar -cvvf '%s' -C %s/temp ./ > /dev/null 2>&1 && rm -rf %s/temp > /dev/null 2>&1", TempFile, TempFileDir, TempFileDir);
It seems pointless to ask for verbose output and then redirect it all to /dev/null.
Gerv
| 1.0 | wget_agent uses -v switch to tar (twice) and then throws all the output away - ---
Author Name: **Gervase Markham**
Original Redmine Issue: 7335, http://www.fossology.org/issues/7335
Original Date: 2014/07/23
---
gerv@hare:/usr/src/fossology$ ack "cvvf"
src/wget_agent/agent/wget_agent.c
404: snprintf(CMD,MAXCMD-1, "tar -cvvf '%s' -C '%s' ./ >/dev/null 2>&1", TempFile, TempFilePath);
505: snprintf(command,MAXCMD-1, "tar -cvvf '%s' -C '%s' ./ >/dev/null 2>&1", GlobalTempFile, TempFileDirectory);
669: snprintf(CMD,MAXCMD-1, "tar -cvvf '%s' -C '%s' ./ %s >/dev/null 2>&1", TempFile, Path, GlobalParam);
689: snprintf(CMD,MAXCMD-1, "tar -cvvf '%s' -C %s/temp ./ > /dev/null 2>&1 && rm -rf %s/temp > /dev/null 2>&1", TempFile, TempFileDir, TempFileDir);
It seems pointless to ask for verbose output and then redirect it all to /dev/null.
Gerv
| non_infrastructure | wget agent uses v switch to tar twice and then throws all the output away author name gervase markham original redmine issue original date gerv hare usr src fossology ack cvvf src wget agent agent wget agent c snprintf cmd maxcmd tar cvvf s c s dev null tempfile tempfilepath snprintf command maxcmd tar cvvf s c s dev null globaltempfile tempfiledirectory snprintf cmd maxcmd tar cvvf s c s s dev null tempfile path globalparam snprintf cmd maxcmd tar cvvf s c s temp dev null rm rf s temp dev null tempfile tempfiledir tempfiledir it seems pointless to ask for verbose output and then redirect it all to dev null gerv | 0 |
139,934 | 20,983,671,508 | IssuesEvent | 2022-03-28 23:08:40 | frnsys/half_earth | https://api.github.com/repos/frnsys/half_earth | closed | Show small faction icon on seat highlight | design | Very minor detail, but when you tap on a parliament seat to see who controls it we highlight that faction's card. But you usually have to scroll to see it which makes it a little unhelpful. We could just show a small icon or even the faction's head instead? | 1.0 | Show small faction icon on seat highlight - Very minor detail, but when you tap on a parliament seat to see who controls it we highlight that faction's card. But you usually have to scroll to see it which makes it a little unhelpful. We could just show a small icon or even the faction's head instead? | non_infrastructure | show small faction icon on seat highlight very minor detail but when you tap on a parliament seat to see who controls it we highlight that faction s card but you usually have to scroll to see it which makes it a little unhelpful we could just show a small icon or even the faction s head instead | 0 |
100,332 | 16,489,859,913 | IssuesEvent | 2021-05-25 01:02:02 | billmcchesney1/goalert | https://api.github.com/repos/billmcchesney1/goalert | opened | CVE-2021-23337 (High) detected in lodash-4.17.20.tgz | security vulnerability | ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p>
<p>Path to dependency file: goalert/web/src/package.json</p>
<p>Path to vulnerable library: goalert/web/src/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- :x: **lodash-4.17.20.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.20","packageFilePaths":["/web/src/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:4.17.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-23337 (High) detected in lodash-4.17.20.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p>
<p>Path to dependency file: goalert/web/src/package.json</p>
<p>Path to vulnerable library: goalert/web/src/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- :x: **lodash-4.17.20.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.20","packageFilePaths":["/web/src/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:4.17.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file goalert web src package json path to vulnerable library goalert web src node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree lodash isminimumfixversionavailable true minimumfixversion lodash basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to command injection via the template function vulnerabilityurl | 0 |
29,946 | 11,787,492,548 | IssuesEvent | 2020-03-17 14:07:42 | adrijshikhar/spotify-player | https://api.github.com/repos/adrijshikhar/spotify-player | opened | CVE-2018-11698 (High) detected in node-sass-4.13.1.tgz, node-sass-v4.13.1 | security vulnerability | ## CVE-2018-11698 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.13.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/spotify-player/package.json</p>
<p>Path to vulnerable library: /spotify-player/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.13.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/adrijshikhar/spotify-player/commit/172ce42afc892e366a594479a5108477592b3bac">172ce42afc892e366a594479a5108477592b3bac</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11698">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11698</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-11698 (High) detected in node-sass-4.13.1.tgz, node-sass-v4.13.1 - ## CVE-2018-11698 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.13.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/spotify-player/package.json</p>
<p>Path to vulnerable library: /spotify-player/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.13.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/adrijshikhar/spotify-player/commit/172ce42afc892e366a594479a5108477592b3bac">172ce42afc892e366a594479a5108477592b3bac</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11698">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11698</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in node sass tgz node sass cve high severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm spotify player package json path to vulnerable library spotify player node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
1,986 | 3,452,365,324 | IssuesEvent | 2015-12-17 03:35:12 | bobandbetty/assassins | https://api.github.com/repos/bobandbetty/assassins | closed | Choose Server Side Framework | infrastructure | Need to choose a framework for the server. Leaning Django. Possibly Flask or Ruby on Rails | 1.0 | Choose Server Side Framework - Need to choose a framework for the server. Leaning Django. Possibly Flask or Ruby on Rails | infrastructure | choose server side framework need to choose a framework for the server leaning django possibly flask or ruby on rails | 1 |
229,794 | 25,375,544,751 | IssuesEvent | 2022-11-21 13:50:58 | stryker-mutator/stryker-net | https://api.github.com/repos/stryker-mutator/stryker-net | closed | stryker.3.0.1.nupkg: 1 vulnerabilities (highest severity is: 5.9) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>stryker.3.0.1.nupkg</b></p></summary>
<p></p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.7.0/system.security.cryptography.xml.4.7.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/stryker-mutator/stryker-net/commit/14d6a329a895287ec783558845ea7fe4f673e59b">14d6a329a895287ec783558845ea7fe4f673e59b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (stryker.3.0.1.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-34716](https://www.mend.io/vulnerability-database/CVE-2022-34716) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | system.security.cryptography.xml.4.7.0.nupkg | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-34716</summary>
### Vulnerable Library - <b>system.security.cryptography.xml.4.7.0.nupkg</b></p>
<p>Provides classes to support the creation and validation of XML digital signatures. The classes in th...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.security.cryptography.xml.4.7.0.nupkg">https://api.nuget.org/packages/system.security.cryptography.xml.4.7.0.nupkg</a></p>
<p>Path to dependency file: /src/Stryker.CLI/Stryker.CLI/Stryker.CLI.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.7.0/system.security.cryptography.xml.4.7.0.nupkg</p>
<p>
Dependency Hierarchy:
- stryker.3.0.1.nupkg (Root Library)
- buildalyzer.4.1.6.nupkg
- microsoft.build.tasks.core.17.0.0.nupkg
- :x: **system.security.cryptography.xml.4.7.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/stryker-mutator/stryker-net/commit/14d6a329a895287ec783558845ea7fe4f673e59b">14d6a329a895287ec783558845ea7fe4f673e59b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Microsoft is releasing this security advisory to provide information about a vulnerability in .NET Core 3.1 and .NET 6.0. An information disclosure vulnerability exists in .NET Core 3.1 and .NET 6.0 that could lead to unauthorized access of privileged information.
## Affected software
* Any .NET 6.0 application running on .NET 6.0.7 or earlier.
* Any .NET Core 3.1 applicaiton running on .NET Core 3.1.27 or earlier.
## Patches
* If you're using .NET 6.0, you should download and install Runtime 6.0.8 or SDK 6.0.108 (for Visual Studio 2022 v17.1) from https://dotnet.microsoft.com/download/dotnet-core/6.0.
* If you're using .NET Core 3.1, you should download and install Runtime 3.1.28 (for Visual Studio 2019 v16.9) from https://dotnet.microsoft.com/download/dotnet-core/3.1.
<p>Publish Date: 2022-08-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-34716>CVE-2022-34716</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-2m65-m22p-9wjw">https://github.com/advisories/GHSA-2m65-m22p-9wjw</a></p>
<p>Release Date: 2022-08-09</p>
<p>Fix Resolution: Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.osx-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.28,6.0.8;System.Security.Cryptography.Xml - 4.7.1,6.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | stryker.3.0.1.nupkg: 1 vulnerabilities (highest severity is: 5.9) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>stryker.3.0.1.nupkg</b></p></summary>
<p></p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.7.0/system.security.cryptography.xml.4.7.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/stryker-mutator/stryker-net/commit/14d6a329a895287ec783558845ea7fe4f673e59b">14d6a329a895287ec783558845ea7fe4f673e59b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (stryker.3.0.1.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-34716](https://www.mend.io/vulnerability-database/CVE-2022-34716) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | system.security.cryptography.xml.4.7.0.nupkg | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-34716</summary>
### Vulnerable Library - <b>system.security.cryptography.xml.4.7.0.nupkg</b></p>
<p>Provides classes to support the creation and validation of XML digital signatures. The classes in th...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.security.cryptography.xml.4.7.0.nupkg">https://api.nuget.org/packages/system.security.cryptography.xml.4.7.0.nupkg</a></p>
<p>Path to dependency file: /src/Stryker.CLI/Stryker.CLI/Stryker.CLI.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.7.0/system.security.cryptography.xml.4.7.0.nupkg</p>
<p>
Dependency Hierarchy:
- stryker.3.0.1.nupkg (Root Library)
- buildalyzer.4.1.6.nupkg
- microsoft.build.tasks.core.17.0.0.nupkg
- :x: **system.security.cryptography.xml.4.7.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/stryker-mutator/stryker-net/commit/14d6a329a895287ec783558845ea7fe4f673e59b">14d6a329a895287ec783558845ea7fe4f673e59b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Microsoft is releasing this security advisory to provide information about a vulnerability in .NET Core 3.1 and .NET 6.0. An information disclosure vulnerability exists in .NET Core 3.1 and .NET 6.0 that could lead to unauthorized access of privileged information.
## Affected software
* Any .NET 6.0 application running on .NET 6.0.7 or earlier.
* Any .NET Core 3.1 applicaiton running on .NET Core 3.1.27 or earlier.
## Patches
* If you're using .NET 6.0, you should download and install Runtime 6.0.8 or SDK 6.0.108 (for Visual Studio 2022 v17.1) from https://dotnet.microsoft.com/download/dotnet-core/6.0.
* If you're using .NET Core 3.1, you should download and install Runtime 3.1.28 (for Visual Studio 2019 v16.9) from https://dotnet.microsoft.com/download/dotnet-core/3.1.
<p>Publish Date: 2022-08-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-34716>CVE-2022-34716</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-2m65-m22p-9wjw">https://github.com/advisories/GHSA-2m65-m22p-9wjw</a></p>
<p>Release Date: 2022-08-09</p>
<p>Fix Resolution: Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.osx-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.28,6.0.8;System.Security.Cryptography.Xml - 4.7.1,6.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_infrastructure | stryker nupkg vulnerabilities highest severity is autoclosed vulnerable library stryker nupkg path to vulnerable library home wss scanner nuget packages system security cryptography xml system security cryptography xml nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in stryker nupkg version remediation available medium system security cryptography xml nupkg transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library system security cryptography xml nupkg provides classes to support the creation and validation of xml digital signatures the classes in th library home page a href path to dependency file src stryker cli stryker cli stryker cli csproj path to vulnerable library home wss scanner nuget packages system security cryptography xml system security cryptography xml nupkg dependency hierarchy stryker nupkg root library buildalyzer nupkg microsoft build tasks core nupkg x system security cryptography xml nupkg vulnerable library found in head commit a href found in base branch master vulnerability details microsoft is releasing this security advisory to provide information about a vulnerability in net core and net an information disclosure vulnerability exists in net core and net that could lead to unauthorized access of privileged information affected software any net application running on net or earlier any net core applicaiton running on net core or earlier patches if you re using net you should download and install runtime or sdk for visual studio from if you re using net core you should download and install runtime for visual studio from publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore app runtime linux arm microsoft aspnetcore app runtime linux microsoft aspnetcore app runtime linux musl arm microsoft aspnetcore app runtime linux musl microsoft aspnetcore app runtime linux musl microsoft aspnetcore app runtime linux microsoft aspnetcore app runtime osx microsoft aspnetcore app runtime win arm microsoft aspnetcore app runtime win microsoft aspnetcore app runtime win microsoft aspnetcore app runtime win system security cryptography xml step up your open source security game with mend | 0 |
33,620 | 27,629,706,320 | IssuesEvent | 2023-03-10 09:55:16 | garritfra/garrit.xyz | https://api.github.com/repos/garritfra/garrit.xyz | closed | Debugging ECS Tasks | post tag:100DaysToOffload tag:infrastructure tag:aws tag:guide tag:note tag:terraform | I just had to debug an application on AWS ECS. The whole procedure is documented in more detail in the [documentation](https://docs.aws.amazon.com/AmazonECS/latest/developerguide/ecs-exec.html), but I think it's beneficial (both for my future self and hopefully to someone out there) to write down the proccess in my own words.
First of all, you need access to the cluster via the [CLI](https://aws.amazon.com/de/cli/). In addition to the CLI, you need the [AWS Session Manager plugin for the CLI](https://docs.aws.amazon.com/systems-manager/latest/userguide/session-manager-working-with-install-plugin.html). If you're on MacOS, you can install that via [Homebrew](https://formulae.brew.sh/cask/session-manager-plugin):
```
brew install --cask session-manager-plugin
```
Next, you need to allow the task you want to debug to be able to execute commands. Since I'm using Terraform, this was just a manner of adding the [`enable_execute_command`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/ecs_service#enable_execute_command) attribute to the service:
```tf
resource "aws_ecs_service" "my_service" {
name = "my-service"
cluster = aws_ecs_cluster.my_cluster.id
task_definition = aws_ecs_task_definition.my_task_definition.id
desired_count = var.app_count
launch_type = "FARGATE"
enable_execute_command = true # TODO: Disable after debugging
}
```
You may also need specify an execution role in the task definition:
```tf
resource "aws_ecs_task_definition" "my_task_definition" {
family = "my-task"
task_role_arn = aws_iam_role.ecs_task_execution_role.arn
execution_role_arn = aws_iam_role.ecs_task_execution_role.arn # <-- Add this
```
Make sure that this role has the correct access rights. There's a nice [troubleshooting guide](https://aws.amazon.com/de/premiumsupport/knowledge-center/ecs-error-execute-command/) going over the required permissions.
If you had to do some modifications, make sure to roll out a new deployment with the fresh settings:
```
aws ecs update-service --cluster my-cluster --service my-service --force-new-deployment
```
Now, you should be able to issue commands against any running container!
```
aws ecs execute-command --cluster westfalen --task <task-id-or-arn> --container my-container --interactive --command="/bin/sh"
```
I hope this helps!
---
This is post 055 of [#100DaysToOffload](https://100daystooffload.com/). | 1.0 | Debugging ECS Tasks - I just had to debug an application on AWS ECS. The whole procedure is documented in more detail in the [documentation](https://docs.aws.amazon.com/AmazonECS/latest/developerguide/ecs-exec.html), but I think it's beneficial (both for my future self and hopefully to someone out there) to write down the proccess in my own words.
First of all, you need access to the cluster via the [CLI](https://aws.amazon.com/de/cli/). In addition to the CLI, you need the [AWS Session Manager plugin for the CLI](https://docs.aws.amazon.com/systems-manager/latest/userguide/session-manager-working-with-install-plugin.html). If you're on MacOS, you can install that via [Homebrew](https://formulae.brew.sh/cask/session-manager-plugin):
```
brew install --cask session-manager-plugin
```
Next, you need to allow the task you want to debug to be able to execute commands. Since I'm using Terraform, this was just a manner of adding the [`enable_execute_command`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/ecs_service#enable_execute_command) attribute to the service:
```tf
resource "aws_ecs_service" "my_service" {
name = "my-service"
cluster = aws_ecs_cluster.my_cluster.id
task_definition = aws_ecs_task_definition.my_task_definition.id
desired_count = var.app_count
launch_type = "FARGATE"
enable_execute_command = true # TODO: Disable after debugging
}
```
You may also need specify an execution role in the task definition:
```tf
resource "aws_ecs_task_definition" "my_task_definition" {
family = "my-task"
task_role_arn = aws_iam_role.ecs_task_execution_role.arn
execution_role_arn = aws_iam_role.ecs_task_execution_role.arn # <-- Add this
```
Make sure that this role has the correct access rights. There's a nice [troubleshooting guide](https://aws.amazon.com/de/premiumsupport/knowledge-center/ecs-error-execute-command/) going over the required permissions.
If you had to do some modifications, make sure to roll out a new deployment with the fresh settings:
```
aws ecs update-service --cluster my-cluster --service my-service --force-new-deployment
```
Now, you should be able to issue commands against any running container!
```
aws ecs execute-command --cluster westfalen --task <task-id-or-arn> --container my-container --interactive --command="/bin/sh"
```
I hope this helps!
---
This is post 055 of [#100DaysToOffload](https://100daystooffload.com/). | infrastructure | debugging ecs tasks i just had to debug an application on aws ecs the whole procedure is documented in more detail in the but i think it s beneficial both for my future self and hopefully to someone out there to write down the proccess in my own words first of all you need access to the cluster via the in addition to the cli you need the if you re on macos you can install that via brew install cask session manager plugin next you need to allow the task you want to debug to be able to execute commands since i m using terraform this was just a manner of adding the attribute to the service tf resource aws ecs service my service name my service cluster aws ecs cluster my cluster id task definition aws ecs task definition my task definition id desired count var app count launch type fargate enable execute command true todo disable after debugging you may also need specify an execution role in the task definition tf resource aws ecs task definition my task definition family my task task role arn aws iam role ecs task execution role arn execution role arn aws iam role ecs task execution role arn add this make sure that this role has the correct access rights there s a nice going over the required permissions if you had to do some modifications make sure to roll out a new deployment with the fresh settings aws ecs update service cluster my cluster service my service force new deployment now you should be able to issue commands against any running container aws ecs execute command cluster westfalen task container my container interactive command bin sh i hope this helps this is post of | 1 |
42,421 | 17,141,152,920 | IssuesEvent | 2021-07-13 09:45:23 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | aws_lb_listener_rule is being validated even when count is zero | needs-triage service/elbv2 | Getting error creating `aws_lb_listener_rule` resource when `count = 0`:
```
Attribute supports 1 item minimum, config has 0 declared
```
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritise this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritise the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
Terraform v0.14.8
AWS provider 3.33.0
### Affected Resource(s)
* aws_lb_listener_rule
### Terraform Configuration Files
```hcl
resource "aws_lb_listener_rule" "path" {
count = var.listener_rule.paths == [] ? 0 : 1
listener_arn = var.listener_rule.listener
action {
type = "forward"
target_group_arn = aws_lb_target_group.this.arn
}
condition {
path_pattern {
values = var.listener_rule.paths
}
}
}
``` | 1.0 | aws_lb_listener_rule is being validated even when count is zero - Getting error creating `aws_lb_listener_rule` resource when `count = 0`:
```
Attribute supports 1 item minimum, config has 0 declared
```
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritise this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritise the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
Terraform v0.14.8
AWS provider 3.33.0
### Affected Resource(s)
* aws_lb_listener_rule
### Terraform Configuration Files
```hcl
resource "aws_lb_listener_rule" "path" {
count = var.listener_rule.paths == [] ? 0 : 1
listener_arn = var.listener_rule.listener
action {
type = "forward"
target_group_arn = aws_lb_target_group.this.arn
}
condition {
path_pattern {
values = var.listener_rule.paths
}
}
}
``` | non_infrastructure | aws lb listener rule is being validated even when count is zero getting error creating aws lb listener rule resource when count attribute supports item minimum config has declared community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritise this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritise the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform aws provider version terraform aws provider affected resource s aws lb listener rule terraform configuration files hcl resource aws lb listener rule path count var listener rule paths listener arn var listener rule listener action type forward target group arn aws lb target group this arn condition path pattern values var listener rule paths | 0 |
332,416 | 10,095,430,505 | IssuesEvent | 2019-07-27 08:47:59 | storybookjs/storybook | https://api.github.com/repos/storybookjs/storybook | closed | Core: Rename `load` to `configure` | high priority maintenance | To support Component Story Format (CSF) in Storybook 5.2 we added a `load` function to all frameworks. This load function uses the `configure` function internally but accepts more arguments (a `require.context` `req` or an array of them) for convenience. It also processes CSF exports from those `reqs`, or allows you to pass a loader function that returns an array of exports.
During CSF feedback, it was pointed out that `load` is basically a superset of `configure` and we don't need to add a new API. This issue is to address that suggestion.
- [x] Rename `load` to `configure`
- [x] Update exports so the new `configure` is exported
- [x] If user passes a configure function (i.e. they are using the old API), be really strict about the return values, and issue a warning with instructions if the values don't match what we expect (either void or an array of objects that contain a `default` export).
- [ ] Update docs and Technical Preview doc
- [ ] Export a deprecated `load`, just for people who are using the beta and don't want to be broken | 1.0 | Core: Rename `load` to `configure` - To support Component Story Format (CSF) in Storybook 5.2 we added a `load` function to all frameworks. This load function uses the `configure` function internally but accepts more arguments (a `require.context` `req` or an array of them) for convenience. It also processes CSF exports from those `reqs`, or allows you to pass a loader function that returns an array of exports.
During CSF feedback, it was pointed out that `load` is basically a superset of `configure` and we don't need to add a new API. This issue is to address that suggestion.
- [x] Rename `load` to `configure`
- [x] Update exports so the new `configure` is exported
- [x] If user passes a configure function (i.e. they are using the old API), be really strict about the return values, and issue a warning with instructions if the values don't match what we expect (either void or an array of objects that contain a `default` export).
- [ ] Update docs and Technical Preview doc
- [ ] Export a deprecated `load`, just for people who are using the beta and don't want to be broken | non_infrastructure | core rename load to configure to support component story format csf in storybook we added a load function to all frameworks this load function uses the configure function internally but accepts more arguments a require context req or an array of them for convenience it also processes csf exports from those reqs or allows you to pass a loader function that returns an array of exports during csf feedback it was pointed out that load is basically a superset of configure and we don t need to add a new api this issue is to address that suggestion rename load to configure update exports so the new configure is exported if user passes a configure function i e they are using the old api be really strict about the return values and issue a warning with instructions if the values don t match what we expect either void or an array of objects that contain a default export update docs and technical preview doc export a deprecated load just for people who are using the beta and don t want to be broken | 0 |
37,759 | 18,760,518,360 | IssuesEvent | 2021-11-05 15:56:00 | dbt-labs/dbt-core | https://api.github.com/repos/dbt-labs/dbt-core | closed | Wasted time spent in `incorporate_indirect_nodes` during node selection | performance | While reviewing #4104, I noticed a lot of wasted time spent in `incorporate_indirect_nodes`.
Here's 3s spent at runtime in the performance testing project:
<img width="1393" alt="Screenshot 2021-11-05 at 08 38 36" src="https://user-images.githubusercontent.com/13897643/140476038-a882dcab-9862-41f0-82b4-925a42e1fc48.png">
This method is frequently unnecessary, _especially_ with the change to the default indirect selection mode in that PR. This method only takes effect when there are nodes selected "indirectly" but not "directly," and we want to incorporate them on the basis of their parents' direct selection.
This one is completely on me: I added this way back in #3235, a change we're mostly turning off by default in v1. I think it should only take a few lines to skip this method when it isn't needed. | True | Wasted time spent in `incorporate_indirect_nodes` during node selection - While reviewing #4104, I noticed a lot of wasted time spent in `incorporate_indirect_nodes`.
Here's 3s spent at runtime in the performance testing project:
<img width="1393" alt="Screenshot 2021-11-05 at 08 38 36" src="https://user-images.githubusercontent.com/13897643/140476038-a882dcab-9862-41f0-82b4-925a42e1fc48.png">
This method is frequently unnecessary, _especially_ with the change to the default indirect selection mode in that PR. This method only takes effect when there are nodes selected "indirectly" but not "directly," and we want to incorporate them on the basis of their parents' direct selection.
This one is completely on me: I added this way back in #3235, a change we're mostly turning off by default in v1. I think it should only take a few lines to skip this method when it isn't needed. | non_infrastructure | wasted time spent in incorporate indirect nodes during node selection while reviewing i noticed a lot of wasted time spent in incorporate indirect nodes here s spent at runtime in the performance testing project img width alt screenshot at src this method is frequently unnecessary especially with the change to the default indirect selection mode in that pr this method only takes effect when there are nodes selected indirectly but not directly and we want to incorporate them on the basis of their parents direct selection this one is completely on me i added this way back in a change we re mostly turning off by default in i think it should only take a few lines to skip this method when it isn t needed | 0 |
237,458 | 7,760,602,941 | IssuesEvent | 2018-06-01 06:51:21 | brave/browser-laptop | https://api.github.com/repos/brave/browser-laptop | closed | Enable Regional Adblocking to match Android | priority/P4 | We'd like to have regional adblocking enabled just like Android.
The specific regional adblockers which are enabled by default in Android should be reflected in the desktop. | 1.0 | Enable Regional Adblocking to match Android - We'd like to have regional adblocking enabled just like Android.
The specific regional adblockers which are enabled by default in Android should be reflected in the desktop. | non_infrastructure | enable regional adblocking to match android we d like to have regional adblocking enabled just like android the specific regional adblockers which are enabled by default in android should be reflected in the desktop | 0 |
32,342 | 26,639,903,725 | IssuesEvent | 2023-01-25 03:01:27 | getodk/central-backend | https://api.github.com/repos/getodk/central-backend | opened | Move off of using Knex for migrations | infrastructure refactor | We currently use Knex for migrations. At one point, we used Knex for all things related to the database. However, in v1.2, we moved to Slonik for everything except migrations. Mostly I think it's been fine to continue to use Knex for migrations, though it's required us to toggle between two pretty different SQL libraries. However, we're also on an old version of Knex, and we know that there are breaking changes on the way to the latest version.
Slonik doesn't have its own framework for migrations, at least when I last checked. However, @alxndrsn made the point that we could probably replicate the important parts of Knex's migration framework ourselves without too much difficulty. We could then rewrite previous migrations, converting from Knex to Slonik or postgres.js.
However, one challenge is that we don't have tests of most migrations, so it would be hard to verify these changes. We also still have high-level questions about how to approach migration tests (#466).
One idea that we've had is to take a snapshot of the schema of a particular past version of Central (version X), then remove all migrations from before that version. Users on a version of Central older than version X would need to upgrade to version X before upgrading to the latest version. If we took this approach, we wouldn't need to rewrite/convert migrations older than version X or write tests for them.
To facilitate this strategy, we're going to try to use as little Knex as possible going forward. That is, we should write as much raw SQL as possible: for example, `db.raw('INSERT ...')`, not `db.insert(...)`. Writing raw SQL is what we usually do outside migrations, so using more raw SQL in migrations should make it easier when we start rewriting migrations newer than version X.
We should also try to test new migrations going forward and continue thinking about #466. | 1.0 | Move off of using Knex for migrations - We currently use Knex for migrations. At one point, we used Knex for all things related to the database. However, in v1.2, we moved to Slonik for everything except migrations. Mostly I think it's been fine to continue to use Knex for migrations, though it's required us to toggle between two pretty different SQL libraries. However, we're also on an old version of Knex, and we know that there are breaking changes on the way to the latest version.
Slonik doesn't have its own framework for migrations, at least when I last checked. However, @alxndrsn made the point that we could probably replicate the important parts of Knex's migration framework ourselves without too much difficulty. We could then rewrite previous migrations, converting from Knex to Slonik or postgres.js.
However, one challenge is that we don't have tests of most migrations, so it would be hard to verify these changes. We also still have high-level questions about how to approach migration tests (#466).
One idea that we've had is to take a snapshot of the schema of a particular past version of Central (version X), then remove all migrations from before that version. Users on a version of Central older than version X would need to upgrade to version X before upgrading to the latest version. If we took this approach, we wouldn't need to rewrite/convert migrations older than version X or write tests for them.
To facilitate this strategy, we're going to try to use as little Knex as possible going forward. That is, we should write as much raw SQL as possible: for example, `db.raw('INSERT ...')`, not `db.insert(...)`. Writing raw SQL is what we usually do outside migrations, so using more raw SQL in migrations should make it easier when we start rewriting migrations newer than version X.
We should also try to test new migrations going forward and continue thinking about #466. | infrastructure | move off of using knex for migrations we currently use knex for migrations at one point we used knex for all things related to the database however in we moved to slonik for everything except migrations mostly i think it s been fine to continue to use knex for migrations though it s required us to toggle between two pretty different sql libraries however we re also on an old version of knex and we know that there are breaking changes on the way to the latest version slonik doesn t have its own framework for migrations at least when i last checked however alxndrsn made the point that we could probably replicate the important parts of knex s migration framework ourselves without too much difficulty we could then rewrite previous migrations converting from knex to slonik or postgres js however one challenge is that we don t have tests of most migrations so it would be hard to verify these changes we also still have high level questions about how to approach migration tests one idea that we ve had is to take a snapshot of the schema of a particular past version of central version x then remove all migrations from before that version users on a version of central older than version x would need to upgrade to version x before upgrading to the latest version if we took this approach we wouldn t need to rewrite convert migrations older than version x or write tests for them to facilitate this strategy we re going to try to use as little knex as possible going forward that is we should write as much raw sql as possible for example db raw insert not db insert writing raw sql is what we usually do outside migrations so using more raw sql in migrations should make it easier when we start rewriting migrations newer than version x we should also try to test new migrations going forward and continue thinking about | 1 |
106,376 | 9,126,597,790 | IssuesEvent | 2019-02-24 22:46:23 | coin-or-tools/BuildTools | https://api.github.com/repos/coin-or-tools/BuildTools | closed | make doxydoc doesn't obey VPATH build | bug configuration tests minor | Issue created by migration from Trac.
Original creator: @LouHafer
Original creation time: 2006-11-02 21:46:13
Assignee: @andrea5w
Version: 0.5
Keywords: VPATH, doxydoc
Once I've configured for a VPATH build, if I run '`make doxydoc`' in the build directory, the `doxydoc/html` directory is created in the distribution directory. At a glance, the fix probably requires a `doxygen.conf.in` file. | 1.0 | make doxydoc doesn't obey VPATH build - Issue created by migration from Trac.
Original creator: @LouHafer
Original creation time: 2006-11-02 21:46:13
Assignee: @andrea5w
Version: 0.5
Keywords: VPATH, doxydoc
Once I've configured for a VPATH build, if I run '`make doxydoc`' in the build directory, the `doxydoc/html` directory is created in the distribution directory. At a glance, the fix probably requires a `doxygen.conf.in` file. | non_infrastructure | make doxydoc doesn t obey vpath build issue created by migration from trac original creator louhafer original creation time assignee version keywords vpath doxydoc once i ve configured for a vpath build if i run make doxydoc in the build directory the doxydoc html directory is created in the distribution directory at a glance the fix probably requires a doxygen conf in file | 0 |
27,439 | 21,711,279,396 | IssuesEvent | 2022-05-10 14:04:17 | sunpy/ndcube | https://api.github.com/repos/sunpy/ndcube | closed | DeprecationWarning: distutils Version classes are deprecated. Use packaging.version instead. | bug help wanted Infrastructure | ### Describe the bug
```
.../ndcube/version.py:27
.../ndcube/version.py:27: DeprecationWarning: distutils Version classes are deprecated. Use packaging.version instead.
for j, piece in enumerate(LooseVersion(version).version[:3]):
```
### To Reproduce
`from ndcube import version`
### What happened?
distutils Version classes are deprecated. Use packaging.version instead.
### Expected behavior
No warning.
### Screenshots
_No response_
### System Details
_No response_
### Installation method
pip | 1.0 | DeprecationWarning: distutils Version classes are deprecated. Use packaging.version instead. - ### Describe the bug
```
.../ndcube/version.py:27
.../ndcube/version.py:27: DeprecationWarning: distutils Version classes are deprecated. Use packaging.version instead.
for j, piece in enumerate(LooseVersion(version).version[:3]):
```
### To Reproduce
`from ndcube import version`
### What happened?
distutils Version classes are deprecated. Use packaging.version instead.
### Expected behavior
No warning.
### Screenshots
_No response_
### System Details
_No response_
### Installation method
pip | infrastructure | deprecationwarning distutils version classes are deprecated use packaging version instead describe the bug ndcube version py ndcube version py deprecationwarning distutils version classes are deprecated use packaging version instead for j piece in enumerate looseversion version version to reproduce from ndcube import version what happened distutils version classes are deprecated use packaging version instead expected behavior no warning screenshots no response system details no response installation method pip | 1 |
31,377 | 25,598,177,194 | IssuesEvent | 2022-12-01 17:47:30 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | closed | Follow up replacements to consistently use "main dashboard" and "entity dashboard" terminology | P2 Good First Issue Rollover Type: Infrastructure | Follow-up to #4161: During approval of that issue, a few follow-up occurrences of the old terms were found in the codebase that should also be updated.
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* CSS classes using `googlesitekit-dashboard-single-url` should have that part replaced with `googlesitekit-entity-dashboard`, in both React and Sass files.
* The one Sass file with that name should also be renamed accordingly.
* The fixture file and reference names used in `tests/e2e/specs/modules/search-console/dashboard-date-range.test.js` should also be renamed to use `main-dashboard` / `mainDashboard` instead of "dashboard" and `entity-dashboard` / `entityDashboard` instead of "dashboard details".
## Implementation Brief
* Using `assets/js/components/DashboardEntityApp.js` and `assets/sass/components/dashboard/_googlesitekit-dashboard-single-url.scss`,
* Replace occurrences of `googlesitekit-dashboard-single-url__*` by `googlesitekit-entity-dashboard__*`.
* Rename `assets/sass/components/dashboard/_googlesitekit-dashboard-single-url.scss` to `assets/sass/components/dashboard/_googlesitekit-entity-dashboard.scss`.
* Using `tests/e2e/specs/modules/search-console/dashboard-date-range.test.js`, make the updates as per the second bullet point in the AC.
### Test Coverage
* No new tests to be added.
## QA Brief
* There are no user facing changes here. However, the main and entity dashboards should be smoke tested to verify they work (especially their styling) as before.
## Changelog entry
* N/A.
| 1.0 | Follow up replacements to consistently use "main dashboard" and "entity dashboard" terminology - Follow-up to #4161: During approval of that issue, a few follow-up occurrences of the old terms were found in the codebase that should also be updated.
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* CSS classes using `googlesitekit-dashboard-single-url` should have that part replaced with `googlesitekit-entity-dashboard`, in both React and Sass files.
* The one Sass file with that name should also be renamed accordingly.
* The fixture file and reference names used in `tests/e2e/specs/modules/search-console/dashboard-date-range.test.js` should also be renamed to use `main-dashboard` / `mainDashboard` instead of "dashboard" and `entity-dashboard` / `entityDashboard` instead of "dashboard details".
## Implementation Brief
* Using `assets/js/components/DashboardEntityApp.js` and `assets/sass/components/dashboard/_googlesitekit-dashboard-single-url.scss`,
* Replace occurrences of `googlesitekit-dashboard-single-url__*` by `googlesitekit-entity-dashboard__*`.
* Rename `assets/sass/components/dashboard/_googlesitekit-dashboard-single-url.scss` to `assets/sass/components/dashboard/_googlesitekit-entity-dashboard.scss`.
* Using `tests/e2e/specs/modules/search-console/dashboard-date-range.test.js`, make the updates as per the second bullet point in the AC.
### Test Coverage
* No new tests to be added.
## QA Brief
* There are no user facing changes here. However, the main and entity dashboards should be smoke tested to verify they work (especially their styling) as before.
## Changelog entry
* N/A.
| infrastructure | follow up replacements to consistently use main dashboard and entity dashboard terminology follow up to during approval of that issue a few follow up occurrences of the old terms were found in the codebase that should also be updated do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria css classes using googlesitekit dashboard single url should have that part replaced with googlesitekit entity dashboard in both react and sass files the one sass file with that name should also be renamed accordingly the fixture file and reference names used in tests specs modules search console dashboard date range test js should also be renamed to use main dashboard maindashboard instead of dashboard and entity dashboard entitydashboard instead of dashboard details implementation brief using assets js components dashboardentityapp js and assets sass components dashboard googlesitekit dashboard single url scss replace occurrences of googlesitekit dashboard single url by googlesitekit entity dashboard rename assets sass components dashboard googlesitekit dashboard single url scss to assets sass components dashboard googlesitekit entity dashboard scss using tests specs modules search console dashboard date range test js make the updates as per the second bullet point in the ac test coverage no new tests to be added qa brief there are no user facing changes here however the main and entity dashboards should be smoke tested to verify they work especially their styling as before changelog entry n a | 1 |
390,208 | 26,852,761,219 | IssuesEvent | 2023-02-03 12:23:13 | ebu/ebu-adm-toolbox | https://api.github.com/repos/ebu/ebu-adm-toolbox | opened | move development guidelines to documentation | documentation | - adjust to match what we actually do, rather than what we were planning
- add instructions for running tests | 1.0 | move development guidelines to documentation - - adjust to match what we actually do, rather than what we were planning
- add instructions for running tests | non_infrastructure | move development guidelines to documentation adjust to match what we actually do rather than what we were planning add instructions for running tests | 0 |
33,862 | 27,956,337,646 | IssuesEvent | 2023-03-24 12:41:42 | aadictive/aadictive.github.io | https://api.github.com/repos/aadictive/aadictive.github.io | closed | Look at the results of the daily master - develop sync github action and clean up the test actions | infrastructure | - [x] Is the sync master-develop github action working as intended (Sync only if there are updates on master branch)?
- [x] Clean up the other workflow file that says test | 1.0 | Look at the results of the daily master - develop sync github action and clean up the test actions - - [x] Is the sync master-develop github action working as intended (Sync only if there are updates on master branch)?
- [x] Clean up the other workflow file that says test | infrastructure | look at the results of the daily master develop sync github action and clean up the test actions is the sync master develop github action working as intended sync only if there are updates on master branch clean up the other workflow file that says test | 1 |
10,182 | 8,406,037,915 | IssuesEvent | 2018-10-11 16:46:49 | 22arw/megaphone | https://api.github.com/repos/22arw/megaphone | closed | Build Authentication Flow | Critical Infrastructure | This is a big part that I will be working on, involving the following items:
- [x] Backend authentication routes (api routes)
- [x] Frontend authentication routes, basically a way to make calls to the backend for authentication.
- [x] The landing screen, Login/Register screen, and the Authenticated Home screen.
The goal is to create a basic auth flow for now where someone can create an account, login, and logout. | 1.0 | Build Authentication Flow - This is a big part that I will be working on, involving the following items:
- [x] Backend authentication routes (api routes)
- [x] Frontend authentication routes, basically a way to make calls to the backend for authentication.
- [x] The landing screen, Login/Register screen, and the Authenticated Home screen.
The goal is to create a basic auth flow for now where someone can create an account, login, and logout. | infrastructure | build authentication flow this is a big part that i will be working on involving the following items backend authentication routes api routes frontend authentication routes basically a way to make calls to the backend for authentication the landing screen login register screen and the authenticated home screen the goal is to create a basic auth flow for now where someone can create an account login and logout | 1 |
9,121 | 7,834,602,328 | IssuesEvent | 2018-06-16 16:05:13 | killbill/killbill | https://api.github.com/repos/killbill/killbill | closed | Generic entity sort order | ENTITLEMENT ext-candidate infrastructure wontfix | `DefaultBlockingState` and `DefaultBlockingStateDao` both implement the same ordering. This could be generalized:
- Make `EntityModelDao` and `Entity` extend `Comparable<BlockingState>`
- Expose the `recordId` (or `totalOrdering`) in the `Entity` objects
| 1.0 | Generic entity sort order - `DefaultBlockingState` and `DefaultBlockingStateDao` both implement the same ordering. This could be generalized:
- Make `EntityModelDao` and `Entity` extend `Comparable<BlockingState>`
- Expose the `recordId` (or `totalOrdering`) in the `Entity` objects
| infrastructure | generic entity sort order defaultblockingstate and defaultblockingstatedao both implement the same ordering this could be generalized make entitymodeldao and entity extend comparable expose the recordid or totalordering in the entity objects | 1 |
142,721 | 11,490,889,030 | IssuesEvent | 2020-02-11 17:54:53 | pokt-network/pocket-core | https://api.github.com/repos/pokt-network/pocket-core | closed | Fix TestMintValidatorAwards | test | TestMintValidatorAwards is failing. This is caused by the Reward Cut off | 1.0 | Fix TestMintValidatorAwards - TestMintValidatorAwards is failing. This is caused by the Reward Cut off | non_infrastructure | fix testmintvalidatorawards testmintvalidatorawards is failing this is caused by the reward cut off | 0 |
43,937 | 13,041,243,114 | IssuesEvent | 2020-07-28 20:00:47 | jgeraigery/nimbus-core-1 | https://api.github.com/repos/jgeraigery/nimbus-core-1 | opened | CVE-2020-13692 (High) detected in postgresql-42.2.5.jar | security vulnerability | ## CVE-2020-13692 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.2.5.jar</b></p></summary>
<p>Java JDBC 4.2 (JRE 8+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc">https://github.com/pgjdbc/pgjdbc</a></p>
<p>Path to dependency file: /tmp/ws-scm/nimbus-core-1/aws/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/org/postgresql/postgresql/42.2.5/postgresql-42.2.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-42.2.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/nimbus-core-1/commit/79321436d185e564fc070879ede6afc398a6ba28">79321436d185e564fc070879ede6afc398a6ba28</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.
<p>Publish Date: 2020-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692>CVE-2020-13692</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13">https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13</a></p>
<p>Release Date: 2020-06-04</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.13</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"42.2.5","isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:42.2.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.13"}],"vulnerabilityIdentifier":"CVE-2020-13692","vulnerabilityDetails":"PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-13692 (High) detected in postgresql-42.2.5.jar - ## CVE-2020-13692 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.2.5.jar</b></p></summary>
<p>Java JDBC 4.2 (JRE 8+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc">https://github.com/pgjdbc/pgjdbc</a></p>
<p>Path to dependency file: /tmp/ws-scm/nimbus-core-1/aws/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/org/postgresql/postgresql/42.2.5/postgresql-42.2.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-42.2.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/nimbus-core-1/commit/79321436d185e564fc070879ede6afc398a6ba28">79321436d185e564fc070879ede6afc398a6ba28</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.
<p>Publish Date: 2020-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692>CVE-2020-13692</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13">https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13</a></p>
<p>Release Date: 2020-06-04</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.13</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"42.2.5","isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:42.2.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.13"}],"vulnerabilityIdentifier":"CVE-2020-13692","vulnerabilityDetails":"PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in postgresql jar cve high severity vulnerability vulnerable library postgresql jar java jdbc jre driver for postgresql database library home page a href path to dependency file tmp ws scm nimbus core aws pom xml path to vulnerable library canner repository org postgresql postgresql postgresql jar dependency hierarchy x postgresql jar vulnerable library found in head commit a href vulnerability details postgresql jdbc driver aka pgjdbc before allows xxe publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org postgresql postgresql rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails postgresql jdbc driver aka pgjdbc before allows xxe vulnerabilityurl | 0 |
7,571 | 6,995,089,783 | IssuesEvent | 2017-12-15 17:46:59 | angular/material2 | https://api.github.com/repos/angular/material2 | closed | Mat-icon : broken links for Material icons font | docs has pr infrastructure P3: important | #### Bug, feature request, or proposal:
Bug
#### What is the expected behavior?
References to the Material icons font should link to the correct URL :
- on the Overview tab, in the **Font icons with ligatures** paragraph
- on the API tab, in the **Directives** paragraph
#### What is the current behavior?
Current links does not link to the correct URL.
#### Is there anything else we should know?
The text from the API tab is correct (http://google.github.io/material-design-icons/#icon-font-for-the-web) but the link is not (https://material.angular.io/components/icon/api#icon-font-for-the-web).
FYI, this is the same erroneous link that is used on the Overview tab. | 1.0 | Mat-icon : broken links for Material icons font - #### Bug, feature request, or proposal:
Bug
#### What is the expected behavior?
References to the Material icons font should link to the correct URL :
- on the Overview tab, in the **Font icons with ligatures** paragraph
- on the API tab, in the **Directives** paragraph
#### What is the current behavior?
Current links does not link to the correct URL.
#### Is there anything else we should know?
The text from the API tab is correct (http://google.github.io/material-design-icons/#icon-font-for-the-web) but the link is not (https://material.angular.io/components/icon/api#icon-font-for-the-web).
FYI, this is the same erroneous link that is used on the Overview tab. | infrastructure | mat icon broken links for material icons font bug feature request or proposal bug what is the expected behavior references to the material icons font should link to the correct url on the overview tab in the font icons with ligatures paragraph on the api tab in the directives paragraph what is the current behavior current links does not link to the correct url is there anything else we should know the text from the api tab is correct but the link is not fyi this is the same erroneous link that is used on the overview tab | 1 |
279,399 | 24,222,460,966 | IssuesEvent | 2022-09-26 12:02:24 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Wazuh 4.3 - SCA policies manual tests - Windows 11 | feed/sca type/test/manual team/threatintel release test/4.3.6 | # Description
This issue aims to manually test the correct functioning of the SCA checks, report the results, and open new issues for any encountered errors.
OS: Windows 11 Enterprise ([OVA](https://developer.microsoft.com/en-us/windows/downloads/virtual-machines/))
SCA policy file: [cis_win11_enterprise.yml](https://github.com/wazuh/wazuh/blob/13087-Win11-CIS-SCA/ruleset/sca/windows/cis_win11_enterprise.yml)
## Tests
For each check in the SCA policy checks:
- The **title**, **description**, **rationale**, and **remediation** must correspond to the ones in the check of the corresponding CIS benchmark file found in https://downloads.cisecurity.org/#/.
- The check command found in the CIS benchmark file should work as expected and must match the rule specified in the check of the `yml` file.
https://github.com/wazuh/wazuh/blob/2bc887c6b8eb751ee0558b4330f4d8b38cae2a17/ruleset/sca/windows/cis_win11_enterprise.yml#L13-L6018
The installers must also be tested:
- Check the scan is executed automatically when installing.
- Check the scan result.
## Test report procedure
All test results must have one of the following statuses:
| | |
|---------------------------------|--------------------------------------------|
| :green_circle: | Everything in the CIS check is correct (description, rule, ID, etc). |
| :red_circle: | There is an error in the CIS check. |
| :yellow_circle: | There is a typo or minor error in the CIS check. |
Any failing test must be properly addressed with a new issue, detailing the error and the possible cause.
Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details.
# Conclusions
All tests have been executed and the results can be found in the issue updates.
## Checks summary
**Total checks:** `395` (26000 - 26394)
All SCA checks have been reviewed and the results can be seen in this issue's updates:
- **1. Account Policies:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1164132336
- **2. Local Policies:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1164279837
- **5. System Services:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1165562733
- **9. Windows Defender Firewall with Advanced Security:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1167143909
- **17. Advanced Audit Policy Configuration:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1168371409
- **18. Administrative Templates (Computer):** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1168507495
## Findings classification
| CIS IDs | STATUS | COMMENTS |
|--|--|--|
|1.1.2, 2.3.4.1, 2.3.7.2, 2.3.9.1, 2.3.10.1, 2.3.10.7, 2.3.10.8, 2.3.11.4, 2.3.14.1, 18.1.2.2, 18.8.3.1, 18.8.5.6, 18.8.22.1.2, 18.8.31.1, 18.8.34.6.5, 18.9.15.1, 18.9.17.1, 18.9.17.2, 18.9.47.4.1, 18.9.47.4.2, 18.9.47.5.1.1, 18.9.47.5.1.2, 18.9.65.3.10.1, 18.9.91.1, 18.9.100.1, 18.9.102.2.3, 18.9.102.2.4, 18.9.105.2.1, 18.9.108.1.1, 18.9.108.4.2 | :red_circle: | Wrong title, description, rationale and/or remediation. |
| 1.1.5, 1.2.1, 1.2.2, 1.2.3, 2.3.4.1, 2.3.7.3, 18.6.1, 18.6.2, 18.6.3, 18.8.3.1, 18.8.5.2, 18.8.5.6, 18.8.7.2, 18.9.17.7, 18.9.47.6.1, 18.9.47.9.1, 18.9.47.9.2, 18.9.47.9.3, 18.9.47.9.4, 18.9.47.16, 18.9.65.3.3.1, 18.9.65.3.9.3, 18.9.65.3.9.5, 18.9.85.2.2, 18.9.100.1, 18.9.108.2.3, 18.9.108.4.1 | :red_circle: | Wrong rule (regex, register path, command or value). |
| 18.9.47.11.11, 18.9.60.1, 18.9.82.1 | :red_circle: | Wrong CIS ID |
| 18.9.100.2 | :red_circle: | Missing ID in the yml (`26375`). |
|1.1.6, 5.17, 5.25, 5.38, 5.40, 18.5.14.1, 18.9.47.16, 18.9.65.3.9.4 | :yellow_circle: | Typos, missing spaces, inconsistencies (for example, L2 is specified in some titles and not in others). |
## Scan
### Logs
As soon as the agent connects, the SCA scan starts automatically as can be seen in these logs:
```
2022/07/04 07:15:49 sca: INFO: Starting evaluation of policy: 'C:\Program Files (x86)\ossec-agent\ruleset\sca\cis_win11_enterprise.yml'
2022/07/04 07:15:50 rootcheck: INFO: Starting rootcheck scan.
2022/07/04 07:15:50 wazuh-agent: INFO: (6000): Starting daemon...
2022/07/04 07:15:50 wazuh-modulesd:syscollector: INFO: Evaluation finished.
2022/07/04 07:15:50 wazuh-agent: INFO: (6010): File integrity monitoring scan frequency: 43200 seconds
2022/07/04 07:15:50 wazuh-agent: INFO: (6008): File integrity monitoring scan started.
2022/07/04 07:15:56 rootcheck: INFO: Ending rootcheck scan.
2022/07/04 07:16:42 wazuh-agent: INFO: (6009): File integrity monitoring scan ended.
2022/07/04 07:16:42 wazuh-agent: INFO: (6012): Real-time file integrity monitoring started.
2022/07/04 07:17:21 sca: INFO: Evaluation finished for policy 'C:\Program Files (x86)\ossec-agent\ruleset\sca\cis_win11_enterprise.yml'
2022/07/04 07:17:21 sca: INFO: Starting evaluation of policy: 'C:\Program Files (x86)\ossec-agent\ruleset\sca\sca_win_audit.yml'
2022/07/04 07:17:23 sca: INFO: Evaluation finished for policy 'C:\Program Files (x86)\ossec-agent\ruleset\sca\sca_win_audit.yml'
2022/07/04 07:17:23 sca: INFO: Security Configuration Assessment scan finished. Duration: 94 seconds.
```
### Result
This is the scan result on the Windows 11 Enterprise OVA which can be found [here](https://developer.microsoft.com/en-us/windows/downloads/virtual-machines/).

| | PASS | NOT APPLICABLE | FAILED |
|--|--|--|--|
|**EXPECTED**| 57 | 258 | 78 |
|**ACTUAL**| 55 | 257 | 82 |
Missmatches are:
- **26006 (1.2.1):** Expected `pass`, actual `fail` (bad regex).
- **26008 (1.2.3):** Expected `pass`, actual `fail` (bad regex).
- **26223 (18.8.3.1):** Expected `not applicable`, actual `fail` (the rule in the yml is different to the rule when the scan is run).
- **26231 (18.8.5.6):** Expected `not applicable`, actual `fail` (the rule in the yml is different to the rule when the scan is run).
- **26232 (18.8.7.2):** Expected `not applicable`, actual `fail` (the rule in the yml is different to the rule when the scan is run).
---------
I also run the scan after changing many of the configuration options according to the remediation. As it can be seen, the changes take effect and the number of checks passed is higher:

| 2.0 | Wazuh 4.3 - SCA policies manual tests - Windows 11 - # Description
This issue aims to manually test the correct functioning of the SCA checks, report the results, and open new issues for any encountered errors.
OS: Windows 11 Enterprise ([OVA](https://developer.microsoft.com/en-us/windows/downloads/virtual-machines/))
SCA policy file: [cis_win11_enterprise.yml](https://github.com/wazuh/wazuh/blob/13087-Win11-CIS-SCA/ruleset/sca/windows/cis_win11_enterprise.yml)
## Tests
For each check in the SCA policy checks:
- The **title**, **description**, **rationale**, and **remediation** must correspond to the ones in the check of the corresponding CIS benchmark file found in https://downloads.cisecurity.org/#/.
- The check command found in the CIS benchmark file should work as expected and must match the rule specified in the check of the `yml` file.
https://github.com/wazuh/wazuh/blob/2bc887c6b8eb751ee0558b4330f4d8b38cae2a17/ruleset/sca/windows/cis_win11_enterprise.yml#L13-L6018
The installers must also be tested:
- Check the scan is executed automatically when installing.
- Check the scan result.
## Test report procedure
All test results must have one of the following statuses:
| | |
|---------------------------------|--------------------------------------------|
| :green_circle: | Everything in the CIS check is correct (description, rule, ID, etc). |
| :red_circle: | There is an error in the CIS check. |
| :yellow_circle: | There is a typo or minor error in the CIS check. |
Any failing test must be properly addressed with a new issue, detailing the error and the possible cause.
Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details.
# Conclusions
All tests have been executed and the results can be found in the issue updates.
## Checks summary
**Total checks:** `395` (26000 - 26394)
All SCA checks have been reviewed and the results can be seen in this issue's updates:
- **1. Account Policies:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1164132336
- **2. Local Policies:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1164279837
- **5. System Services:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1165562733
- **9. Windows Defender Firewall with Advanced Security:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1167143909
- **17. Advanced Audit Policy Configuration:** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1168371409
- **18. Administrative Templates (Computer):** https://github.com/wazuh/wazuh/issues/13961#issuecomment-1168507495
## Findings classification
| CIS IDs | STATUS | COMMENTS |
|--|--|--|
|1.1.2, 2.3.4.1, 2.3.7.2, 2.3.9.1, 2.3.10.1, 2.3.10.7, 2.3.10.8, 2.3.11.4, 2.3.14.1, 18.1.2.2, 18.8.3.1, 18.8.5.6, 18.8.22.1.2, 18.8.31.1, 18.8.34.6.5, 18.9.15.1, 18.9.17.1, 18.9.17.2, 18.9.47.4.1, 18.9.47.4.2, 18.9.47.5.1.1, 18.9.47.5.1.2, 18.9.65.3.10.1, 18.9.91.1, 18.9.100.1, 18.9.102.2.3, 18.9.102.2.4, 18.9.105.2.1, 18.9.108.1.1, 18.9.108.4.2 | :red_circle: | Wrong title, description, rationale and/or remediation. |
| 1.1.5, 1.2.1, 1.2.2, 1.2.3, 2.3.4.1, 2.3.7.3, 18.6.1, 18.6.2, 18.6.3, 18.8.3.1, 18.8.5.2, 18.8.5.6, 18.8.7.2, 18.9.17.7, 18.9.47.6.1, 18.9.47.9.1, 18.9.47.9.2, 18.9.47.9.3, 18.9.47.9.4, 18.9.47.16, 18.9.65.3.3.1, 18.9.65.3.9.3, 18.9.65.3.9.5, 18.9.85.2.2, 18.9.100.1, 18.9.108.2.3, 18.9.108.4.1 | :red_circle: | Wrong rule (regex, register path, command or value). |
| 18.9.47.11.11, 18.9.60.1, 18.9.82.1 | :red_circle: | Wrong CIS ID |
| 18.9.100.2 | :red_circle: | Missing ID in the yml (`26375`). |
|1.1.6, 5.17, 5.25, 5.38, 5.40, 18.5.14.1, 18.9.47.16, 18.9.65.3.9.4 | :yellow_circle: | Typos, missing spaces, inconsistencies (for example, L2 is specified in some titles and not in others). |
## Scan
### Logs
As soon as the agent connects, the SCA scan starts automatically as can be seen in these logs:
```
2022/07/04 07:15:49 sca: INFO: Starting evaluation of policy: 'C:\Program Files (x86)\ossec-agent\ruleset\sca\cis_win11_enterprise.yml'
2022/07/04 07:15:50 rootcheck: INFO: Starting rootcheck scan.
2022/07/04 07:15:50 wazuh-agent: INFO: (6000): Starting daemon...
2022/07/04 07:15:50 wazuh-modulesd:syscollector: INFO: Evaluation finished.
2022/07/04 07:15:50 wazuh-agent: INFO: (6010): File integrity monitoring scan frequency: 43200 seconds
2022/07/04 07:15:50 wazuh-agent: INFO: (6008): File integrity monitoring scan started.
2022/07/04 07:15:56 rootcheck: INFO: Ending rootcheck scan.
2022/07/04 07:16:42 wazuh-agent: INFO: (6009): File integrity monitoring scan ended.
2022/07/04 07:16:42 wazuh-agent: INFO: (6012): Real-time file integrity monitoring started.
2022/07/04 07:17:21 sca: INFO: Evaluation finished for policy 'C:\Program Files (x86)\ossec-agent\ruleset\sca\cis_win11_enterprise.yml'
2022/07/04 07:17:21 sca: INFO: Starting evaluation of policy: 'C:\Program Files (x86)\ossec-agent\ruleset\sca\sca_win_audit.yml'
2022/07/04 07:17:23 sca: INFO: Evaluation finished for policy 'C:\Program Files (x86)\ossec-agent\ruleset\sca\sca_win_audit.yml'
2022/07/04 07:17:23 sca: INFO: Security Configuration Assessment scan finished. Duration: 94 seconds.
```
### Result
This is the scan result on the Windows 11 Enterprise OVA which can be found [here](https://developer.microsoft.com/en-us/windows/downloads/virtual-machines/).

| | PASS | NOT APPLICABLE | FAILED |
|--|--|--|--|
|**EXPECTED**| 57 | 258 | 78 |
|**ACTUAL**| 55 | 257 | 82 |
Missmatches are:
- **26006 (1.2.1):** Expected `pass`, actual `fail` (bad regex).
- **26008 (1.2.3):** Expected `pass`, actual `fail` (bad regex).
- **26223 (18.8.3.1):** Expected `not applicable`, actual `fail` (the rule in the yml is different to the rule when the scan is run).
- **26231 (18.8.5.6):** Expected `not applicable`, actual `fail` (the rule in the yml is different to the rule when the scan is run).
- **26232 (18.8.7.2):** Expected `not applicable`, actual `fail` (the rule in the yml is different to the rule when the scan is run).
---------
I also run the scan after changing many of the configuration options according to the remediation. As it can be seen, the changes take effect and the number of checks passed is higher:

| non_infrastructure | wazuh sca policies manual tests windows description this issue aims to manually test the correct functioning of the sca checks report the results and open new issues for any encountered errors os windows enterprise sca policy file tests for each check in the sca policy checks the title description rationale and remediation must correspond to the ones in the check of the corresponding cis benchmark file found in the check command found in the cis benchmark file should work as expected and must match the rule specified in the check of the yml file the installers must also be tested check the scan is executed automatically when installing check the scan result test report procedure all test results must have one of the following statuses green circle everything in the cis check is correct description rule id etc red circle there is an error in the cis check yellow circle there is a typo or minor error in the cis check any failing test must be properly addressed with a new issue detailing the error and the possible cause please attach any documents screenshots or tables to the issue update with the results this report can be used by the auditors to dig deeper into any possible failures and details conclusions all tests have been executed and the results can be found in the issue updates checks summary total checks all sca checks have been reviewed and the results can be seen in this issue s updates account policies local policies system services windows defender firewall with advanced security advanced audit policy configuration administrative templates computer findings classification cis ids status comments red circle wrong title description rationale and or remediation red circle wrong rule regex register path command or value red circle wrong cis id red circle missing id in the yml yellow circle typos missing spaces inconsistencies for example is specified in some titles and not in others scan logs as soon as the agent connects the sca scan starts automatically as can be seen in these logs sca info starting evaluation of policy c program files ossec agent ruleset sca cis enterprise yml rootcheck info starting rootcheck scan wazuh agent info starting daemon wazuh modulesd syscollector info evaluation finished wazuh agent info file integrity monitoring scan frequency seconds wazuh agent info file integrity monitoring scan started rootcheck info ending rootcheck scan wazuh agent info file integrity monitoring scan ended wazuh agent info real time file integrity monitoring started sca info evaluation finished for policy c program files ossec agent ruleset sca cis enterprise yml sca info starting evaluation of policy c program files ossec agent ruleset sca sca win audit yml sca info evaluation finished for policy c program files ossec agent ruleset sca sca win audit yml sca info security configuration assessment scan finished duration seconds result this is the scan result on the windows enterprise ova which can be found pass not applicable failed expected actual missmatches are expected pass actual fail bad regex expected pass actual fail bad regex expected not applicable actual fail the rule in the yml is different to the rule when the scan is run expected not applicable actual fail the rule in the yml is different to the rule when the scan is run expected not applicable actual fail the rule in the yml is different to the rule when the scan is run i also run the scan after changing many of the configuration options according to the remediation as it can be seen the changes take effect and the number of checks passed is higher | 0 |
835 | 2,579,722,444 | IssuesEvent | 2015-02-13 12:47:45 | jakobbossek/ecr | https://api.github.com/repos/jakobbossek/ecr | closed | Overwork and overthink the population objects and functions | enhancement ugly code | Especially ```mergePopulations``` is far from efficient and programmatically ugly. | 1.0 | Overwork and overthink the population objects and functions - Especially ```mergePopulations``` is far from efficient and programmatically ugly. | non_infrastructure | overwork and overthink the population objects and functions especially mergepopulations is far from efficient and programmatically ugly | 0 |
33,995 | 28,074,157,106 | IssuesEvent | 2023-03-29 21:35:28 | kaleidawave/ezno | https://api.github.com/repos/kaleidawave/ezno | closed | Improve CI testing and publishing | infrastructure | ### Testing
- **Run fuzzing tests in parallel**. Will show red light correctly and be faster
- Get cargo fuzz from: https://github.com/rust-fuzz/cargo-fuzz/releases/tag/0.11.2
- Split up testing step by crates (eventually this should mean that it should only run tests for crates that have changed)
- Have some sort of feature matrix for testing
- **Add hyperfine and flamegraph** performance tracing
- Enforce Cargo.toml formats
*need to check whether caching is working correctly*
### Publishing
- Automatically add `ezno-ast-generator` if `ezno-parser` is updated
| 1.0 | Improve CI testing and publishing - ### Testing
- **Run fuzzing tests in parallel**. Will show red light correctly and be faster
- Get cargo fuzz from: https://github.com/rust-fuzz/cargo-fuzz/releases/tag/0.11.2
- Split up testing step by crates (eventually this should mean that it should only run tests for crates that have changed)
- Have some sort of feature matrix for testing
- **Add hyperfine and flamegraph** performance tracing
- Enforce Cargo.toml formats
*need to check whether caching is working correctly*
### Publishing
- Automatically add `ezno-ast-generator` if `ezno-parser` is updated
| infrastructure | improve ci testing and publishing testing run fuzzing tests in parallel will show red light correctly and be faster get cargo fuzz from split up testing step by crates eventually this should mean that it should only run tests for crates that have changed have some sort of feature matrix for testing add hyperfine and flamegraph performance tracing enforce cargo toml formats need to check whether caching is working correctly publishing automatically add ezno ast generator if ezno parser is updated | 1 |
11,432 | 9,195,470,772 | IssuesEvent | 2019-03-07 02:33:30 | HumanCellAtlas/secondary-analysis | https://api.github.com/repos/HumanCellAtlas/secondary-analysis | closed | Move devops code from Lira to secondary-analysis | devops infrastructure 🍿Snack | A lot of work has gone into improving our deployment and they are not specific to only Lira, they should live in the secondary-analysis repo. | 1.0 | Move devops code from Lira to secondary-analysis - A lot of work has gone into improving our deployment and they are not specific to only Lira, they should live in the secondary-analysis repo. | infrastructure | move devops code from lira to secondary analysis a lot of work has gone into improving our deployment and they are not specific to only lira they should live in the secondary analysis repo | 1 |
34,286 | 29,190,172,127 | IssuesEvent | 2023-05-19 19:08:49 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | opened | Investigate E2E Test Execution Performance | type: task tool: monorepo infrastructure | ### Description
Right now we are currently running the entire E2E test suite in a single job. This is absolutely our _slowest_ GitHub check, and as such, provides the biggest win. One of the things that I noticed when working in the Gutenberg repository is that they use [Playwright's Shard feature](https://playwright.dev/docs/test-parallel#shard-tests-between-multiple-machines) in order to spread the testing load across a number of workers. These improvements, and others, likely exist that we can take advantage of to speed up the E2E tests. We should dive into these and figure out what we can do to improve performance.
## Acceptance Criteria
Rather than diving straight in, the goal of this task is to curate a list of specific issues that we can address individually. The output here should be more GitHub Issues of appropriate size to tackle in future cycles. | 1.0 | Investigate E2E Test Execution Performance - ### Description
Right now we are currently running the entire E2E test suite in a single job. This is absolutely our _slowest_ GitHub check, and as such, provides the biggest win. One of the things that I noticed when working in the Gutenberg repository is that they use [Playwright's Shard feature](https://playwright.dev/docs/test-parallel#shard-tests-between-multiple-machines) in order to spread the testing load across a number of workers. These improvements, and others, likely exist that we can take advantage of to speed up the E2E tests. We should dive into these and figure out what we can do to improve performance.
## Acceptance Criteria
Rather than diving straight in, the goal of this task is to curate a list of specific issues that we can address individually. The output here should be more GitHub Issues of appropriate size to tackle in future cycles. | infrastructure | investigate test execution performance description right now we are currently running the entire test suite in a single job this is absolutely our slowest github check and as such provides the biggest win one of the things that i noticed when working in the gutenberg repository is that they use in order to spread the testing load across a number of workers these improvements and others likely exist that we can take advantage of to speed up the tests we should dive into these and figure out what we can do to improve performance acceptance criteria rather than diving straight in the goal of this task is to curate a list of specific issues that we can address individually the output here should be more github issues of appropriate size to tackle in future cycles | 1 |
44,338 | 13,055,239,733 | IssuesEvent | 2020-07-30 01:03:46 | jtimberlake/skf-flask | https://api.github.com/repos/jtimberlake/skf-flask | opened | CVE-2020-8203 (High) detected in lodash-4.17.15.tgz | security vulnerability | ## CVE-2020-8203 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/skf-flask/Angular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/skf-flask/Angular/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-4.1.0.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution attack when using _.zipObjectDeep in lodash <= 4.17.15.
<p>Publish Date: 2020-07-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p>
<p>Release Date: 2020-07-23</p>
<p>Fix Resolution: lodash - 4.17.19</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","isTransitiveDependency":true,"dependencyTree":"karma:4.1.0;lodash:4.17.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.19"}],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash \u003c\u003d 4.17.15.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-8203 (High) detected in lodash-4.17.15.tgz - ## CVE-2020-8203 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/skf-flask/Angular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/skf-flask/Angular/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-4.1.0.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution attack when using _.zipObjectDeep in lodash <= 4.17.15.
<p>Publish Date: 2020-07-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p>
<p>Release Date: 2020-07-23</p>
<p>Fix Resolution: lodash - 4.17.19</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","isTransitiveDependency":true,"dependencyTree":"karma:4.1.0;lodash:4.17.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.19"}],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash \u003c\u003d 4.17.15.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file tmp ws scm skf flask angular package json path to vulnerable library tmp ws scm skf flask angular node modules lodash package json dependency hierarchy karma tgz root library x lodash tgz vulnerable library vulnerability details prototype pollution attack when using zipobjectdeep in lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails prototype pollution attack when using zipobjectdeep in lodash vulnerabilityurl | 0 |
703,229 | 24,150,476,627 | IssuesEvent | 2022-09-21 23:47:38 | googleapis/nodejs-ai-platform | https://api.github.com/repos/googleapis/nodejs-ai-platform | closed | AI platform get video classification model evaluation: should get the evaluation from the specified model failed | type: bug priority: p1 flakybot: issue api: vertex-ai | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: e1c5cd6b5d03afb03911ba9aa685457aa359a602
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/69dc3b23-298d-461c-9279-cd7551a90fc9), [Sponge](http://sponge2/69dc3b23-298d-461c-9279-cd7551a90fc9)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./get-model-evaluation-video-classification.js 8596984660557299712 7092045712224944128 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
Error: Command failed: node ./get-model-evaluation-video-classification.js 8596984660557299712 7092045712224944128 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/get-model-evaluation-video-classification.test.js:24:28)
at Context.<anonymous> (test/get-model-evaluation-video-classification.test.js:34:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | 1.0 | AI platform get video classification model evaluation: should get the evaluation from the specified model failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: e1c5cd6b5d03afb03911ba9aa685457aa359a602
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/69dc3b23-298d-461c-9279-cd7551a90fc9), [Sponge](http://sponge2/69dc3b23-298d-461c-9279-cd7551a90fc9)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./get-model-evaluation-video-classification.js 8596984660557299712 7092045712224944128 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
Error: Command failed: node ./get-model-evaluation-video-classification.js 8596984660557299712 7092045712224944128 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/get-model-evaluation-video-classification.test.js:24:28)
at Context.<anonymous> (test/get-model-evaluation-video-classification.test.js:34:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | non_infrastructure | ai platform get video classification model evaluation should get the evaluation from the specified model failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node get model evaluation video classification js ucaip sample tests us unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error command failed node get model evaluation video classification js ucaip sample tests us unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at checkexecsyncerror child process js at object execsync child process js at execsync test get model evaluation video classification test js at context test get model evaluation video classification test js at processimmediate internal timers js | 0 |
48,024 | 6,065,982,731 | IssuesEvent | 2017-06-14 17:26:48 | chapel-lang/chapel | https://api.github.com/repos/chapel-lang/chapel | closed | Parallel Algorithms using domain maps | area: Language type: Design user issue | Hi
I'm new to Chapel language i find it during my GSoC proposal time. I have a doubt regarding using domain maps in parallel algorithms, "Using domain maps in parallel algorithms, the programmers don't have to worry about their initial placement of data in the global address space" what this sentence means exactly. I have understand what domain map are ? And how to use it. But i didn't find meaning for this sentence clearly. Whether domain is for only linear data structures or it can be used for any user-defined data type ?
Thank You so much | 1.0 | Parallel Algorithms using domain maps - Hi
I'm new to Chapel language i find it during my GSoC proposal time. I have a doubt regarding using domain maps in parallel algorithms, "Using domain maps in parallel algorithms, the programmers don't have to worry about their initial placement of data in the global address space" what this sentence means exactly. I have understand what domain map are ? And how to use it. But i didn't find meaning for this sentence clearly. Whether domain is for only linear data structures or it can be used for any user-defined data type ?
Thank You so much | non_infrastructure | parallel algorithms using domain maps hi i m new to chapel language i find it during my gsoc proposal time i have a doubt regarding using domain maps in parallel algorithms using domain maps in parallel algorithms the programmers don t have to worry about their initial placement of data in the global address space what this sentence means exactly i have understand what domain map are and how to use it but i didn t find meaning for this sentence clearly whether domain is for only linear data structures or it can be used for any user defined data type thank you so much | 0 |
401,644 | 27,334,312,193 | IssuesEvent | 2023-02-26 01:56:12 | typeorm/typeorm | https://api.github.com/repos/typeorm/typeorm | opened | how show sql query in this code ? | documentation requires triage | ### What was unclear or otherwise insufficient?
async findAll() {
const user = await this.userRepository.find({
relations: {
posts: true,
},
});
return user;
}
how show sql query in this code ?
### Recommended Fix
async findAll() {
const user = await this.userRepository.find({
relations: {
posts: true,
},
});
return user;
}
how show sql query in this code ?
### Additional Context
_No response_
### Are you willing to resolve this issue by submitting a Pull Request?
Yes, I have the time, and I know how to start. | 1.0 | how show sql query in this code ? - ### What was unclear or otherwise insufficient?
async findAll() {
const user = await this.userRepository.find({
relations: {
posts: true,
},
});
return user;
}
how show sql query in this code ?
### Recommended Fix
async findAll() {
const user = await this.userRepository.find({
relations: {
posts: true,
},
});
return user;
}
how show sql query in this code ?
### Additional Context
_No response_
### Are you willing to resolve this issue by submitting a Pull Request?
Yes, I have the time, and I know how to start. | non_infrastructure | how show sql query in this code what was unclear or otherwise insufficient async findall const user await this userrepository find relations posts true return user how show sql query in this code recommended fix async findall const user await this userrepository find relations posts true return user how show sql query in this code additional context no response are you willing to resolve this issue by submitting a pull request yes i have the time and i know how to start | 0 |
23,198 | 15,930,847,434 | IssuesEvent | 2021-04-14 01:54:14 | MissouriMRR/IARC-2020 | https://api.github.com/repos/MissouriMRR/IARC-2020 | closed | Pipeline failure flags | Infrastructure Pipeline enhancement feature good first issue important vision | ## Description of Problem
<!--Comment: Phrase from the perspective of a user, also known as a user story-->
In order to differentiate between crash and no bounding boxes being detected, failure flags should be implemented in the pipeline. These flags will signify that an algorithm failed to run to completion.
## Requirements
<!--Comment: Should be generic to any possible implementation, usually in the form of a to-do list-->
- [ ] Add failure flags. These can be implemented as a new separate object type.
- [ ] Flags should be algorithm-specific. For example, each module algorithm (is_in_frame(), get_center(), etc.) should have its own flag.
- [ ] Flag object should be specific to pipeline state. For example, would not show obstacle detection failure flag during module detection state.
- [ ] Flags should be saved per run. Need a queue or similar to store flags every time run is called.
- [ ] Flags should be able to be outputted as string so they can be logged.
| 1.0 | Pipeline failure flags - ## Description of Problem
<!--Comment: Phrase from the perspective of a user, also known as a user story-->
In order to differentiate between crash and no bounding boxes being detected, failure flags should be implemented in the pipeline. These flags will signify that an algorithm failed to run to completion.
## Requirements
<!--Comment: Should be generic to any possible implementation, usually in the form of a to-do list-->
- [ ] Add failure flags. These can be implemented as a new separate object type.
- [ ] Flags should be algorithm-specific. For example, each module algorithm (is_in_frame(), get_center(), etc.) should have its own flag.
- [ ] Flag object should be specific to pipeline state. For example, would not show obstacle detection failure flag during module detection state.
- [ ] Flags should be saved per run. Need a queue or similar to store flags every time run is called.
- [ ] Flags should be able to be outputted as string so they can be logged.
| infrastructure | pipeline failure flags description of problem in order to differentiate between crash and no bounding boxes being detected failure flags should be implemented in the pipeline these flags will signify that an algorithm failed to run to completion requirements add failure flags these can be implemented as a new separate object type flags should be algorithm specific for example each module algorithm is in frame get center etc should have its own flag flag object should be specific to pipeline state for example would not show obstacle detection failure flag during module detection state flags should be saved per run need a queue or similar to store flags every time run is called flags should be able to be outputted as string so they can be logged | 1 |
31,864 | 26,204,958,529 | IssuesEvent | 2023-01-03 21:29:56 | catch22/octave-doctest | https://api.github.com/repos/catch22/octave-doctest | closed | change default branch | area: infrastructure | @catch22 can you change the "default branch" under Settings from `master` to `main`?
On one of my other repos it looks like this:

| 1.0 | change default branch - @catch22 can you change the "default branch" under Settings from `master` to `main`?
On one of my other repos it looks like this:

| infrastructure | change default branch can you change the default branch under settings from master to main on one of my other repos it looks like this | 1 |
139,356 | 5,367,976,733 | IssuesEvent | 2017-02-22 07:00:26 | code-corps/code-corps-ember | https://api.github.com/repos/code-corps/code-corps-ember | closed | Consider extracting out check buttons into an async check button component | medium priority | Should DRY up some similarities between `category-item` and `role-item`.
| 1.0 | Consider extracting out check buttons into an async check button component - Should DRY up some similarities between `category-item` and `role-item`.
| non_infrastructure | consider extracting out check buttons into an async check button component should dry up some similarities between category item and role item | 0 |
161,231 | 25,308,769,658 | IssuesEvent | 2022-11-17 15:54:54 | primefaces/primereact | https://api.github.com/repos/primefaces/primereact | closed | ContextMenu: Trigger via left-click configurable | Resolution: Workaround Resolution: By Design | ### Describe the feature you would like to see added
Please add the option to configure ContextMenu to be triggered via left click
### Is your feature request related to a problem?
The current right click activation makes development really difficult since it interferes with browser inspection, which is also triggered by a right click.
### Describe the solution you'd like
The ability to trigger ContextMenu via left click
### Describe alternatives you have considered
_No response_
### Additional context
_No response_ | 1.0 | ContextMenu: Trigger via left-click configurable - ### Describe the feature you would like to see added
Please add the option to configure ContextMenu to be triggered via left click
### Is your feature request related to a problem?
The current right click activation makes development really difficult since it interferes with browser inspection, which is also triggered by a right click.
### Describe the solution you'd like
The ability to trigger ContextMenu via left click
### Describe alternatives you have considered
_No response_
### Additional context
_No response_ | non_infrastructure | contextmenu trigger via left click configurable describe the feature you would like to see added please add the option to configure contextmenu to be triggered via left click is your feature request related to a problem the current right click activation makes development really difficult since it interferes with browser inspection which is also triggered by a right click describe the solution you d like the ability to trigger contextmenu via left click describe alternatives you have considered no response additional context no response | 0 |
30,440 | 13,246,658,640 | IssuesEvent | 2020-08-19 16:01:04 | appvia/kore | https://api.github.com/repos/appvia/kore | closed | Save schema versions to service objects | EPIC: Service Provisioning | ## What
We want to save what schema was used for the service provider, service and service credentials configurations, so we can properly support migrations | 1.0 | Save schema versions to service objects - ## What
We want to save what schema was used for the service provider, service and service credentials configurations, so we can properly support migrations | non_infrastructure | save schema versions to service objects what we want to save what schema was used for the service provider service and service credentials configurations so we can properly support migrations | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.