Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
30,076 | 24,526,251,642 | IssuesEvent | 2022-10-11 13:21:11 | Altinn/altinn-platform | https://api.github.com/repos/Altinn/altinn-platform | opened | Enable Azure Policy Add-On on apps AKS cluster | kind/user-story solution/apps status/draft ops/infrastructure | ### Description
For more details please visit [Understand Azure Policy for Azure Kubernetes Service](https://docs.microsoft.com/en-ie/azure/governance/policy/concepts/rego-for-aks)
### Additional Information
_No response_
### Tasks
_No response_
### Acceptance Criterias
_No response_ | 1.0 | Enable Azure Policy Add-On on apps AKS cluster - ### Description
For more details please visit [Understand Azure Policy for Azure Kubernetes Service](https://docs.microsoft.com/en-ie/azure/governance/policy/concepts/rego-for-aks)
### Additional Information
_No response_
### Tasks
_No response_
### Acceptance Criterias
_No response_ | infrastructure | enable azure policy add on on apps aks cluster description for more details please visit additional information no response tasks no response acceptance criterias no response | 1 |
94,425 | 11,871,396,886 | IssuesEvent | 2020-03-26 14:20:06 | liqd/adhocracy-plus | https://api.github.com/repos/liqd/adhocracy-plus | opened | autofollow not transparent | Type: UX/UI or design | **URL:**
**user:** registered user
**expected behaviour:** I follow a project only after clicking the follow button
**behaviour:** I automatically follow the project after submitting an idea
**important screensize:**
**device & browser:**
**Comment/Question:** I like the idea behind the autofollow but think we need to be more transparent there, e.g. asking after submitting an idea if user wants to follow project.
Screenshot?
| 1.0 | autofollow not transparent - **URL:**
**user:** registered user
**expected behaviour:** I follow a project only after clicking the follow button
**behaviour:** I automatically follow the project after submitting an idea
**important screensize:**
**device & browser:**
**Comment/Question:** I like the idea behind the autofollow but think we need to be more transparent there, e.g. asking after submitting an idea if user wants to follow project.
Screenshot?
| non_infrastructure | autofollow not transparent url user registered user expected behaviour i follow a project only after clicking the follow button behaviour i automatically follow the project after submitting an idea important screensize device browser comment question i like the idea behind the autofollow but think we need to be more transparent there e g asking after submitting an idea if user wants to follow project screenshot | 0 |
36,253 | 5,040,453,404 | IssuesEvent | 2016-12-19 05:39:35 | RMajewski/InstrumenteSoftware | https://api.github.com/repos/RMajewski/InstrumenteSoftware | closed | TestDataNameDescription | completed test todo | Testet die Datenklasse "DataNameDescription" auf richtige Verarbeitung der Daten:
- ID
- Name
- Beschreibung | 1.0 | TestDataNameDescription - Testet die Datenklasse "DataNameDescription" auf richtige Verarbeitung der Daten:
- ID
- Name
- Beschreibung | non_infrastructure | testdatanamedescription testet die datenklasse datanamedescription auf richtige verarbeitung der daten id name beschreibung | 0 |
157,767 | 24,723,757,599 | IssuesEvent | 2022-10-20 12:42:26 | matrix-org/dendrite | https://api.github.com/repos/matrix-org/dendrite | closed | Dendrite tries to pump 26MB messages over Kafka | design:scaling design:components needs discussion | ### Background information
- **Dendrite version or git SHA**: 0.3.2
- **Monolith or Polylith?**: Monolith
- **SQLite3 or Postgres?**: Postgres
- **Running in Docker?**: Yes
- **`go version`**: Whatever is in the Docker container
- **Kafka version**: Docker: bitnami/kafka:2.6.0
<!--
This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
Dendrite tries to send pretty huge messages over Kafka when trying to join certain rooms.
<!-- Describe here the problem that you are experiencing -->
### Steps to reproduce
* Set up Dendrite in monolith mode with Postgres and Kafka
* Join a few rooms
* Try to join #fedora:matrix.org
* Watch it try to join for ~10 hours and eventually fail with the following error message hidden in the logs;
```
time="2020-12-08T11:24:12.820136945Z" level=error msg="Write to kafka failed" func="WriteOutputEvents\n\t" file=" [github.com/matrix-org/dendrite/roomserver/internal/input/input.go:113]" error="kafka: Failed to produce message to topic DendriteOutputRoomEvent: kafka server: Message was too large, server rejected it to avoid allocation error." message_bytes=26292271
```
<!--
Describe how what happens differs from what you expected.
If you can identify any relevant log snippets from server logs, please include
those (please be careful to remove any personal or private data). Please surround them with
``` (three backticks, on a line on their own), so that they are formatted legibly.
Alternatively, please send logs to @kegan:matrix.org or @neilalexander:matrix.org
with a link to the respective Github issue, thanks!
-->
| 2.0 | Dendrite tries to pump 26MB messages over Kafka - ### Background information
- **Dendrite version or git SHA**: 0.3.2
- **Monolith or Polylith?**: Monolith
- **SQLite3 or Postgres?**: Postgres
- **Running in Docker?**: Yes
- **`go version`**: Whatever is in the Docker container
- **Kafka version**: Docker: bitnami/kafka:2.6.0
<!--
This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
Dendrite tries to send pretty huge messages over Kafka when trying to join certain rooms.
<!-- Describe here the problem that you are experiencing -->
### Steps to reproduce
* Set up Dendrite in monolith mode with Postgres and Kafka
* Join a few rooms
* Try to join #fedora:matrix.org
* Watch it try to join for ~10 hours and eventually fail with the following error message hidden in the logs;
```
time="2020-12-08T11:24:12.820136945Z" level=error msg="Write to kafka failed" func="WriteOutputEvents\n\t" file=" [github.com/matrix-org/dendrite/roomserver/internal/input/input.go:113]" error="kafka: Failed to produce message to topic DendriteOutputRoomEvent: kafka server: Message was too large, server rejected it to avoid allocation error." message_bytes=26292271
```
<!--
Describe how what happens differs from what you expected.
If you can identify any relevant log snippets from server logs, please include
those (please be careful to remove any personal or private data). Please surround them with
``` (three backticks, on a line on their own), so that they are formatted legibly.
Alternatively, please send logs to @kegan:matrix.org or @neilalexander:matrix.org
with a link to the respective Github issue, thanks!
-->
| non_infrastructure | dendrite tries to pump messages over kafka background information dendrite version or git sha monolith or polylith monolith or postgres postgres running in docker yes go version whatever is in the docker container kafka version docker bitnami kafka this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description dendrite tries to send pretty huge messages over kafka when trying to join certain rooms steps to reproduce set up dendrite in monolith mode with postgres and kafka join a few rooms try to join fedora matrix org watch it try to join for hours and eventually fail with the following error message hidden in the logs time level error msg write to kafka failed func writeoutputevents n t file error kafka failed to produce message to topic dendriteoutputroomevent kafka server message was too large server rejected it to avoid allocation error message bytes describe how what happens differs from what you expected if you can identify any relevant log snippets from server logs please include those please be careful to remove any personal or private data please surround them with three backticks on a line on their own so that they are formatted legibly alternatively please send logs to kegan matrix org or neilalexander matrix org with a link to the respective github issue thanks | 0 |
26,095 | 19,661,288,956 | IssuesEvent | 2022-01-10 17:16:30 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | Developer Experience: Publish infrastructure service catalog | operations VSP-Initiative infrastructure | ## Problem Statement
It is difficult for developers to navigate the platform’s infrastructure services. It is also hard to provide quality support when services are not undefined.
How might we make it easier for developers to navigate the platform's available infrastructure services?
How might we provide better documentation for the platform's available infrastructure services?
How might we better support for the platform's available infrastructure services?
## Hypothesis or Bet
We believe that this will result in a higher quality infrastructure support experience for VFS and platform crew teams.
We believe that this will result in a higher quality infrastructure support experience for infrastructure and SRE personnel.
## We will know we're done when... ("Definition of Done")
Five (5) infrastructure services are published to the Platform Website. And, the following exists for each service:
- the service is listed on the platform website
- developer-oriented documentation for the service exists on the platform website
- there is internal documentation (such as a playbook) for platform support personnel
- there are machine-readable GitHub issue templates (ie with boilerplate text, fields, choices; pre-defined by YAML; w/ limited free-form input)
## Known Blockers/Dependencies
TBD
## Projected Launch Date
TBD
## Launch Checklist
### Is this service / tool / feature...
### ... tested?
- [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it.
- [ ] ... and issues discovered in usability testing have been addressed.
* _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._
- [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching
- [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD
### ... documented?
- [ ] New documentation is written pursuant to our [documentation style guide](https://vfs.atlassian.net/wiki/spaces/AP/pages/622264362/Style+guide)
- [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0)
* _List the existing product that this initiative fits within, or add a new product to this list._
- [ ] Internal-facing: there's a [Product Outline](https://vfs.atlassian.net/wiki/spaces/PMCP/pages/1924628490/Product+Outline+Template)
- [ ] External-facing: a [User Guide on Platform Website](https://vfs.atlassian.net/wiki/spaces/AP/pages/1477017691/Platform+website+guidelines) exists for this product/feature tool
- [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings)
### ... measurable
- [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it.
* For help, see: [Analytics team](https://depo-platform-documentation.scrollhelp.site/analytics-monitoring/Analytics-customer-support-guide.1586823275.html)
- [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose).
### When you're ready to launch...
- [ ] Conduct a [go/no-go] (https://vfs.atlassian.net/wiki/spaces/AP/pages/1670938648/Platform+Crew+Office+Hours#Go%2FNo-Go) when you're almost ready to launch.
## Required Artifacts
### Documentation
* **`PRODUCT_NAME`**: _directory name used for your product documentation_
* **Product Outline**: _link to Product Outline_
* **User Guide**: _link to User Guide_
### Testing
* **Usability test**: _link to GitHub issue, or provide reason for skipping_
* **Manual QA**: _link to GitHub issue or documented results_
* **Automated tests**: _link to tests, or "N/A"_
### Measurement
* **Success metrics**: _link to where success metrics are measured, or where they're defined (Product Outline is OK), or provide reason for skipping_
* **Release plan**: _link to Release Plan ticket_
## TODOs
- [ ] Convert this issue to an epic
- [ ] Add your team's label to this epic
| 1.0 | Developer Experience: Publish infrastructure service catalog - ## Problem Statement
It is difficult for developers to navigate the platform’s infrastructure services. It is also hard to provide quality support when services are not undefined.
How might we make it easier for developers to navigate the platform's available infrastructure services?
How might we provide better documentation for the platform's available infrastructure services?
How might we better support for the platform's available infrastructure services?
## Hypothesis or Bet
We believe that this will result in a higher quality infrastructure support experience for VFS and platform crew teams.
We believe that this will result in a higher quality infrastructure support experience for infrastructure and SRE personnel.
## We will know we're done when... ("Definition of Done")
Five (5) infrastructure services are published to the Platform Website. And, the following exists for each service:
- the service is listed on the platform website
- developer-oriented documentation for the service exists on the platform website
- there is internal documentation (such as a playbook) for platform support personnel
- there are machine-readable GitHub issue templates (ie with boilerplate text, fields, choices; pre-defined by YAML; w/ limited free-form input)
## Known Blockers/Dependencies
TBD
## Projected Launch Date
TBD
## Launch Checklist
### Is this service / tool / feature...
### ... tested?
- [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it.
- [ ] ... and issues discovered in usability testing have been addressed.
* _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._
- [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching
- [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD
### ... documented?
- [ ] New documentation is written pursuant to our [documentation style guide](https://vfs.atlassian.net/wiki/spaces/AP/pages/622264362/Style+guide)
- [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0)
* _List the existing product that this initiative fits within, or add a new product to this list._
- [ ] Internal-facing: there's a [Product Outline](https://vfs.atlassian.net/wiki/spaces/PMCP/pages/1924628490/Product+Outline+Template)
- [ ] External-facing: a [User Guide on Platform Website](https://vfs.atlassian.net/wiki/spaces/AP/pages/1477017691/Platform+website+guidelines) exists for this product/feature tool
- [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings)
### ... measurable
- [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it.
* For help, see: [Analytics team](https://depo-platform-documentation.scrollhelp.site/analytics-monitoring/Analytics-customer-support-guide.1586823275.html)
- [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose).
### When you're ready to launch...
- [ ] Conduct a [go/no-go] (https://vfs.atlassian.net/wiki/spaces/AP/pages/1670938648/Platform+Crew+Office+Hours#Go%2FNo-Go) when you're almost ready to launch.
## Required Artifacts
### Documentation
* **`PRODUCT_NAME`**: _directory name used for your product documentation_
* **Product Outline**: _link to Product Outline_
* **User Guide**: _link to User Guide_
### Testing
* **Usability test**: _link to GitHub issue, or provide reason for skipping_
* **Manual QA**: _link to GitHub issue or documented results_
* **Automated tests**: _link to tests, or "N/A"_
### Measurement
* **Success metrics**: _link to where success metrics are measured, or where they're defined (Product Outline is OK), or provide reason for skipping_
* **Release plan**: _link to Release Plan ticket_
## TODOs
- [ ] Convert this issue to an epic
- [ ] Add your team's label to this epic
| infrastructure | developer experience publish infrastructure service catalog problem statement it is difficult for developers to navigate the platform’s infrastructure services it is also hard to provide quality support when services are not undefined how might we make it easier for developers to navigate the platform s available infrastructure services how might we provide better documentation for the platform s available infrastructure services how might we better support for the platform s available infrastructure services hypothesis or bet we believe that this will result in a higher quality infrastructure support experience for vfs and platform crew teams we believe that this will result in a higher quality infrastructure support experience for infrastructure and sre personnel we will know we re done when definition of done five infrastructure services are published to the platform website and the following exists for each service the service is listed on the platform website developer oriented documentation for the service exists on the platform website there is internal documentation such as a playbook for platform support personnel there are machine readable github issue templates ie with boilerplate text fields choices pre defined by yaml w limited free form input known blockers dependencies tbd projected launch date tbd launch checklist is this service tool feature tested usability test todo link has been performed to validate that new changes enable users to do what was intended and that these changes don t worsen quality elsewhere if usability test isn t relevant for this change document the reason for skipping it and issues discovered in usability testing have been addressed note on skipping metrics that show the impact of before after can be a substitute for usability testing end to end or is complete to validate there are no high severity issues before launching if applicable new functionality has thorough automated tests running in ci cd documented new documentation is written pursuant to our product is included in the list the existing product that this initiative fits within or add a new product to this list internal facing there s a external facing a exists for this product feature tool if applicable post to for external communication about this change e g vsp newsletter customer facing meetings measurable if applicable this change has clearly defined success metrics with instrumentation of those analytics where possible or a reason documented for skipping it for help see this change has an accompanying when you re ready to launch conduct a when you re almost ready to launch required artifacts documentation product name directory name used for your product documentation product outline link to product outline user guide link to user guide testing usability test link to github issue or provide reason for skipping manual qa link to github issue or documented results automated tests link to tests or n a measurement success metrics link to where success metrics are measured or where they re defined product outline is ok or provide reason for skipping release plan link to release plan ticket todos convert this issue to an epic add your team s label to this epic | 1 |
460,314 | 13,208,123,413 | IssuesEvent | 2020-08-15 02:32:46 | Stooberton/ACF-3 | https://api.github.com/repos/Stooberton/ACF-3 | closed | [BUG] ACF-2 fuel tanks become disabled on spawn | bug compatibility high priority info needed | **Short Description**
Some fuel tanks from ACF-2 dupes might become disabled when they're spawned because their mass is slightly lower than the minimum legal mass.
| 1.0 | [BUG] ACF-2 fuel tanks become disabled on spawn - **Short Description**
Some fuel tanks from ACF-2 dupes might become disabled when they're spawned because their mass is slightly lower than the minimum legal mass.
| non_infrastructure | acf fuel tanks become disabled on spawn short description some fuel tanks from acf dupes might become disabled when they re spawned because their mass is slightly lower than the minimum legal mass | 0 |
274,027 | 20,823,207,438 | IssuesEvent | 2022-03-18 17:31:14 | phylum-dev/cli | https://api.github.com/repos/phylum-dev/cli | opened | `phylum projects set-thresholds` help output is missing for `name` argument | bug good first issue documentation low complexity needs triage | # Overview
It is not clear how to use the `phylum projects set-thresholds` command, based on system output and help messages. The documentation...online and in the help message itself...should exist and be clear about the intended usage. Right now, users have to guess what the missing `<name>` argument is meant to represent.
## How To Reproduce
Steps to reproduce this behavior:
1. Issue the `phylum projects set-thresholds` command or the `phylum projects set-thresholds --help` command
2. See example output, showing the missing `<name>` help text:
```sh
➜ cli git:(development) target/debug/phylum projects set-thresholds
error: The following required arguments were not provided:
<name>
USAGE:
phylum projects set-thresholds <name>
For more information try --help
➜ cli git:(development) target/debug/phylum projects set-thresholds --help
phylum-projects-set-thresholds
Set risk domain thresholds for a projects
USAGE:
phylum projects set-thresholds <name>
ARGS:
<name>
OPTIONS:
-h, --help Print help information
```
## Expected Behavior
The `<name>` argument has a help message that makes it clear that a project name is expected, where the project names come from the output of the `phylum projects list` command.
## Additional Context
N/A
| 1.0 | `phylum projects set-thresholds` help output is missing for `name` argument - # Overview
It is not clear how to use the `phylum projects set-thresholds` command, based on system output and help messages. The documentation...online and in the help message itself...should exist and be clear about the intended usage. Right now, users have to guess what the missing `<name>` argument is meant to represent.
## How To Reproduce
Steps to reproduce this behavior:
1. Issue the `phylum projects set-thresholds` command or the `phylum projects set-thresholds --help` command
2. See example output, showing the missing `<name>` help text:
```sh
➜ cli git:(development) target/debug/phylum projects set-thresholds
error: The following required arguments were not provided:
<name>
USAGE:
phylum projects set-thresholds <name>
For more information try --help
➜ cli git:(development) target/debug/phylum projects set-thresholds --help
phylum-projects-set-thresholds
Set risk domain thresholds for a projects
USAGE:
phylum projects set-thresholds <name>
ARGS:
<name>
OPTIONS:
-h, --help Print help information
```
## Expected Behavior
The `<name>` argument has a help message that makes it clear that a project name is expected, where the project names come from the output of the `phylum projects list` command.
## Additional Context
N/A
| non_infrastructure | phylum projects set thresholds help output is missing for name argument overview it is not clear how to use the phylum projects set thresholds command based on system output and help messages the documentation online and in the help message itself should exist and be clear about the intended usage right now users have to guess what the missing argument is meant to represent how to reproduce steps to reproduce this behavior issue the phylum projects set thresholds command or the phylum projects set thresholds help command see example output showing the missing help text sh ➜ cli git development target debug phylum projects set thresholds error the following required arguments were not provided usage phylum projects set thresholds for more information try help ➜ cli git development target debug phylum projects set thresholds help phylum projects set thresholds set risk domain thresholds for a projects usage phylum projects set thresholds args options h help print help information expected behavior the argument has a help message that makes it clear that a project name is expected where the project names come from the output of the phylum projects list command additional context n a | 0 |
108,800 | 4,350,772,829 | IssuesEvent | 2016-07-31 13:31:34 | rm-code/On-The-Roadside | https://api.github.com/repos/rm-code/On-The-Roadside | opened | Explosions | Priority: High Status: Accepted Type: Feature | OTR will include explosive weapons at some point in the future (rocket launchers, grenades and so on). The current system only handles projectile weapons and therefore needs to be extended.
This is from an early test:

An easy way to implement explosions for now would be to take all tiles in the radius of the explosion and hit them a certain amount of damage. This would however ignore walls and other objects which could block the blast.
Instead all tiles inside of the radius should be checked by shooting rays at them and only those which are reached by the ray will be hit by the explosion.
| 1.0 | Explosions - OTR will include explosive weapons at some point in the future (rocket launchers, grenades and so on). The current system only handles projectile weapons and therefore needs to be extended.
This is from an early test:

An easy way to implement explosions for now would be to take all tiles in the radius of the explosion and hit them a certain amount of damage. This would however ignore walls and other objects which could block the blast.
Instead all tiles inside of the radius should be checked by shooting rays at them and only those which are reached by the ray will be hit by the explosion.
| non_infrastructure | explosions otr will include explosive weapons at some point in the future rocket launchers grenades and so on the current system only handles projectile weapons and therefore needs to be extended this is from an early test an easy way to implement explosions for now would be to take all tiles in the radius of the explosion and hit them a certain amount of damage this would however ignore walls and other objects which could block the blast instead all tiles inside of the radius should be checked by shooting rays at them and only those which are reached by the ray will be hit by the explosion | 0 |
23,904 | 16,679,892,337 | IssuesEvent | 2021-06-07 21:34:02 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | DNS outage impacting builds | area-infrastructure | <!-- runfo report start -->
Runfo Tracking Issue: [DNS outage ](https://runfo.azurewebsites.net/tracking/issue/157)
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1072106](https://dev.azure.com/dnceng/public/_build/results?buildId=1072106)|[PR 31497](https://github.com/dotnet/aspnetcore/pull/31497)|Test: Templates - Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1055529](https://dev.azure.com/dnceng/public/_build/results?buildId=1055529)|Rolling|Build: Windows ARM|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1055474](https://dev.azure.com/dnceng/public/_build/results?buildId=1055474)|[PR 31199](https://github.com/dotnet/aspnetcore/pull/31199)|Test: Templates - Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1052768](https://dev.azure.com/dnceng/public/_build/results?buildId=1052768)|[PR 31156](https://github.com/dotnet/aspnetcore/pull/31156)|Build: Windows x64/x86|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1052562](https://dev.azure.com/dnceng/public/_build/results?buildId=1052562)|[PR 31153](https://github.com/dotnet/aspnetcore/pull/31153)|Build: Windows ARM64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1048296](https://dev.azure.com/dnceng/public/_build/results?buildId=1048296)|[PR 30907](https://github.com/dotnet/aspnetcore/pull/30907)|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1048296](https://dev.azure.com/dnceng/public/_build/results?buildId=1048296)|[PR 30907](https://github.com/dotnet/aspnetcore/pull/30907)|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1048096](https://dev.azure.com/dnceng/public/_build/results?buildId=1048096)|Rolling|Build: Windows x64/x86|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1047610](https://dev.azure.com/dnceng/public/_build/results?buildId=1047610)|[PR 30874](https://github.com/dotnet/aspnetcore/pull/30874)|Build: Windows x64/x86|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1046361](https://dev.azure.com/dnceng/public/_build/results?buildId=1046361)|Rolling|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1045863](https://dev.azure.com/dnceng/public/_build/results?buildId=1045863)|[PR 31033](https://github.com/dotnet/aspnetcore/pull/31033)|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1043504](https://dev.azure.com/dnceng/public/_build/results?buildId=1043504)|[PR 30999](https://github.com/dotnet/aspnetcore/pull/30999)|Tests: Helix x64|
Build Result Summary
|Day Hit Count|Week Hit Count|Month Hit Count|
|---|---|---|
|0|0|11|
<!-- runfo report end -->
| 1.0 | DNS outage impacting builds - <!-- runfo report start -->
Runfo Tracking Issue: [DNS outage ](https://runfo.azurewebsites.net/tracking/issue/157)
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1072106](https://dev.azure.com/dnceng/public/_build/results?buildId=1072106)|[PR 31497](https://github.com/dotnet/aspnetcore/pull/31497)|Test: Templates - Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1055529](https://dev.azure.com/dnceng/public/_build/results?buildId=1055529)|Rolling|Build: Windows ARM|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1055474](https://dev.azure.com/dnceng/public/_build/results?buildId=1055474)|[PR 31199](https://github.com/dotnet/aspnetcore/pull/31199)|Test: Templates - Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1052768](https://dev.azure.com/dnceng/public/_build/results?buildId=1052768)|[PR 31156](https://github.com/dotnet/aspnetcore/pull/31156)|Build: Windows x64/x86|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1052562](https://dev.azure.com/dnceng/public/_build/results?buildId=1052562)|[PR 31153](https://github.com/dotnet/aspnetcore/pull/31153)|Build: Windows ARM64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1048296](https://dev.azure.com/dnceng/public/_build/results?buildId=1048296)|[PR 30907](https://github.com/dotnet/aspnetcore/pull/30907)|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1048296](https://dev.azure.com/dnceng/public/_build/results?buildId=1048296)|[PR 30907](https://github.com/dotnet/aspnetcore/pull/30907)|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1048096](https://dev.azure.com/dnceng/public/_build/results?buildId=1048096)|Rolling|Build: Windows x64/x86|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1047610](https://dev.azure.com/dnceng/public/_build/results?buildId=1047610)|[PR 30874](https://github.com/dotnet/aspnetcore/pull/30874)|Build: Windows x64/x86|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1046361](https://dev.azure.com/dnceng/public/_build/results?buildId=1046361)|Rolling|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1045863](https://dev.azure.com/dnceng/public/_build/results?buildId=1045863)|[PR 31033](https://github.com/dotnet/aspnetcore/pull/31033)|Test: Windows Server 2016 x64|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[1043504](https://dev.azure.com/dnceng/public/_build/results?buildId=1043504)|[PR 30999](https://github.com/dotnet/aspnetcore/pull/30999)|Tests: Helix x64|
Build Result Summary
|Day Hit Count|Week Hit Count|Month Hit Count|
|---|---|---|
|0|0|11|
<!-- runfo report end -->
| infrastructure | dns outage impacting builds runfo tracking issue definition build kind job name templates windows server windows arm templates windows server windows windows windows server windows server windows windows windows server windows server helix build result summary day hit count week hit count month hit count | 1 |
6,674 | 8,955,150,974 | IssuesEvent | 2019-01-26 04:02:46 | SpongePowered/SpongeForge | https://api.github.com/repos/SpongePowered/SpongeForge | closed | Dimensional Pockets 2 startup crash | system: world gen type: mod incompatibility version: 1.12 | **I am currently running**
<!-- If you don't use the latest version, please tell us why. -->
- SpongeForge version:
- Forge version:
- Java version:
- Operating System:
<!-- Please include ALL mods/plugins you had installed when your issue happened, you can get a list of
your mods and plugins by running "/sponge plugins" and/or "/sponge mods" -->
- Plugins/Mods:
<!-- Please include as much information as possible. For the description, assume we have no idea how
mods work, be as detailed as possible and include a step by step reproduction. It is recommended
you try to reproduce the issue you are having yourself with as few mods as possible. -->
**Issue Description**
<!-- Please provide a *full* server log (and crash-report if applicable).
Go to https://gist.github.com/ and upload them there, then paste the resulting link here!
Don't use hastebin/pastebin or other similar sites, as they have a history of quickly
deleting files before we can look at them. -->With regard to using with Dimensional-Pockets, once this MOD is added, the server can't start!
But if you delete spong or Dimensional-Pockets, then the server will start.
It is emphasized that the two can not be used together, but once it is used, it cannot start.
I begged the author to correct this BUG?
This is the wrong document:
https://gist.github.com/phit/0ac9f7d1e1009cdfcf1e86b3ef8590ea
| True | Dimensional Pockets 2 startup crash - **I am currently running**
<!-- If you don't use the latest version, please tell us why. -->
- SpongeForge version:
- Forge version:
- Java version:
- Operating System:
<!-- Please include ALL mods/plugins you had installed when your issue happened, you can get a list of
your mods and plugins by running "/sponge plugins" and/or "/sponge mods" -->
- Plugins/Mods:
<!-- Please include as much information as possible. For the description, assume we have no idea how
mods work, be as detailed as possible and include a step by step reproduction. It is recommended
you try to reproduce the issue you are having yourself with as few mods as possible. -->
**Issue Description**
<!-- Please provide a *full* server log (and crash-report if applicable).
Go to https://gist.github.com/ and upload them there, then paste the resulting link here!
Don't use hastebin/pastebin or other similar sites, as they have a history of quickly
deleting files before we can look at them. -->With regard to using with Dimensional-Pockets, once this MOD is added, the server can't start!
But if you delete spong or Dimensional-Pockets, then the server will start.
It is emphasized that the two can not be used together, but once it is used, it cannot start.
I begged the author to correct this BUG?
This is the wrong document:
https://gist.github.com/phit/0ac9f7d1e1009cdfcf1e86b3ef8590ea
| non_infrastructure | dimensional pockets startup crash i am currently running spongeforge version forge version java version operating system please include all mods plugins you had installed when your issue happened you can get a list of your mods and plugins by running sponge plugins and or sponge mods plugins mods please include as much information as possible for the description assume we have no idea how mods work be as detailed as possible and include a step by step reproduction it is recommended you try to reproduce the issue you are having yourself with as few mods as possible issue description please provide a full server log and crash report if applicable go to and upload them there then paste the resulting link here don t use hastebin pastebin or other similar sites as they have a history of quickly deleting files before we can look at them with regard to using with dimensional pockets once this mod is added the server can t start but if you delete spong or dimensional pockets then the server will start it is emphasized that the two can not be used together but once it is used it cannot start i begged the author to correct this bug this is the wrong document: | 0 |
345,627 | 10,370,673,944 | IssuesEvent | 2019-09-08 14:34:57 | yalla-coop/death | https://api.github.com/repos/yalla-coop/death | closed | Should be able to delete the This or That questions | bug priority-2 | Currently you can't delete This or That questions, which is included within the scope | 1.0 | Should be able to delete the This or That questions - Currently you can't delete This or That questions, which is included within the scope | non_infrastructure | should be able to delete the this or that questions currently you can t delete this or that questions which is included within the scope | 0 |
13,756 | 10,450,921,018 | IssuesEvent | 2019-09-19 11:42:39 | ampproject/amp.dev | https://api.github.com/repos/ampproject/amp.dev | closed | Make a version of `grow build` that doesn't use GH tokens | Category: Infrastructure P2: Medium | This came out of the discussion at https://github.com/ampproject/docs/pull/1002#issuecomment-398517598. It's needed so that we can run `grow build` as part of PR checks for incoming PRs, and Travis doesn't make secure tokens available to PR builds.
Currently, it's only `scripts/import_docs.js` that uses the token. Since we don't really want to do a deploy duding PR builds, perhaps we can restrict them to just building the docs and making sure nothing broke?
https://github.com/ampproject/docs/blob/3670e95f11318d904c6a8ae5759faf88706007be/scripts/import_docs.js#L13-L16
@pbakaus @bpaduch Thoughts?
| 1.0 | Make a version of `grow build` that doesn't use GH tokens - This came out of the discussion at https://github.com/ampproject/docs/pull/1002#issuecomment-398517598. It's needed so that we can run `grow build` as part of PR checks for incoming PRs, and Travis doesn't make secure tokens available to PR builds.
Currently, it's only `scripts/import_docs.js` that uses the token. Since we don't really want to do a deploy duding PR builds, perhaps we can restrict them to just building the docs and making sure nothing broke?
https://github.com/ampproject/docs/blob/3670e95f11318d904c6a8ae5759faf88706007be/scripts/import_docs.js#L13-L16
@pbakaus @bpaduch Thoughts?
| infrastructure | make a version of grow build that doesn t use gh tokens this came out of the discussion at it s needed so that we can run grow build as part of pr checks for incoming prs and travis doesn t make secure tokens available to pr builds currently it s only scripts import docs js that uses the token since we don t really want to do a deploy duding pr builds perhaps we can restrict them to just building the docs and making sure nothing broke pbakaus bpaduch thoughts | 1 |
19,305 | 13,212,232,190 | IssuesEvent | 2020-08-16 05:35:19 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | jenkins_plugin in check mode always reports changed status | affects_2.8 bot_closed bug collection collection:community.general module needs_collection_redirect needs_triage python3 support:community web_infrastructure | <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
When in check mode, the `jenkins_plugin` module always reports a changed status, even when no change would happen.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`jenkins_plugin` module
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible 2.8.1
config file = /home/dead10ck/src/infra/streamline/ansible/ansible.cfg
configured module search path = ['/home/dead10ck/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python3.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
ANSIBLE_PIPELINING(/home/dead10ck/src/infra/streamline/ansible/ansible.cfg) = True
DEFAULT_HOST_LIST(/home/dead10ck/src/infra/streamline/ansible/ansible.cfg) = ['/home/dead10ck/src/infra/streamline/ansible/inventory/aws_ec2.yml']
HOST_KEY_CHECKING(env: ANSIBLE_HOST_KEY_CHECKING) = False
RETRY_FILES_ENABLED(/home/dead10ck/src/infra/streamline/ansible/ansible.cfg) = False
```
##### OS / ENVIRONMENT
Fedora 30 Workstation
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
1. Use the `jenkins_plugin` module to install a plugin
1. Run it again in check mode. It will report changed status.
1. Run it again without check mode. It will report no change.
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
When in check mode, it should only report a change when a change would happen.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
A change is reported when no change would happen. | 1.0 | jenkins_plugin in check mode always reports changed status - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
When in check mode, the `jenkins_plugin` module always reports a changed status, even when no change would happen.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`jenkins_plugin` module
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible 2.8.1
config file = /home/dead10ck/src/infra/streamline/ansible/ansible.cfg
configured module search path = ['/home/dead10ck/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python3.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
ANSIBLE_PIPELINING(/home/dead10ck/src/infra/streamline/ansible/ansible.cfg) = True
DEFAULT_HOST_LIST(/home/dead10ck/src/infra/streamline/ansible/ansible.cfg) = ['/home/dead10ck/src/infra/streamline/ansible/inventory/aws_ec2.yml']
HOST_KEY_CHECKING(env: ANSIBLE_HOST_KEY_CHECKING) = False
RETRY_FILES_ENABLED(/home/dead10ck/src/infra/streamline/ansible/ansible.cfg) = False
```
##### OS / ENVIRONMENT
Fedora 30 Workstation
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
1. Use the `jenkins_plugin` module to install a plugin
1. Run it again in check mode. It will report changed status.
1. Run it again without check mode. It will report no change.
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
When in check mode, it should only report a change when a change would happen.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
A change is reported when no change would happen. | infrastructure | jenkins plugin in check mode always reports changed status summary when in check mode the jenkins plugin module always reports a changed status even when no change would happen issue type bug report component name jenkins plugin module ansible version paste below ansible config file home src infra streamline ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default may configuration paste below ansible pipelining home src infra streamline ansible ansible cfg true default host list home src infra streamline ansible ansible cfg host key checking env ansible host key checking false retry files enabled home src infra streamline ansible ansible cfg false os environment fedora workstation steps to reproduce use the jenkins plugin module to install a plugin run it again in check mode it will report changed status run it again without check mode it will report no change expected results when in check mode it should only report a change when a change would happen actual results a change is reported when no change would happen | 1 |
30,673 | 24,990,422,942 | IssuesEvent | 2022-11-02 18:16:04 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | InitializeToolset flakiness in correctness builds | Area-Infrastructure untriaged | <!-- runfo report start -->
Runfo Tracking Issue: [Correctness build InitializeToolset issues](https://runfo.azurewebsites.net/tracking/issue/237)
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[66363](https://dev.azure.com/dnceng-public/public/_build/results?buildId=66363)|Rolling|Build_Windows_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Build_Windows_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Build_Windows_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Correctness_Rebuild|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Correctness_Rebuild|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Build_Windows_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[62903](https://dev.azure.com/dnceng-public/public/_build/results?buildId=62903)|[PR 64957](https://github.com/dotnet/roslyn/pull/64957)|Test_Windows_Desktop_Debug_64|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_CoreClr_IOperation_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_64|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_CoreClr_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_CoreClr_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56316](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56316)|[PR 64803](https://github.com/dotnet/roslyn/pull/64803)|Test_Windows_Desktop_Release_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56316](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56316)|[PR 64803](https://github.com/dotnet/roslyn/pull/64803)|Test_Windows_CoreClr_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56269](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56269)|Rolling|Test_Windows_CoreClr_Debug_Single_Machine|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[51811](https://dev.azure.com/dnceng-public/public/_build/results?buildId=51811)|Rolling|Test_Windows_CoreClr_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[39111](https://dev.azure.com/dnceng-public/public/_build/results?buildId=39111)|Rolling|Source-Build (Managed)|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[39111](https://dev.azure.com/dnceng-public/public/_build/results?buildId=39111)|Rolling|Build_Windows_Release|
Build Result Summary
|Day Hit Count|Week Hit Count|Month Hit Count|
|---|---|---|
|0|1|8|
<!-- runfo report end -->
| 1.0 | InitializeToolset flakiness in correctness builds - <!-- runfo report start -->
Runfo Tracking Issue: [Correctness build InitializeToolset issues](https://runfo.azurewebsites.net/tracking/issue/237)
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[66363](https://dev.azure.com/dnceng-public/public/_build/results?buildId=66363)|Rolling|Build_Windows_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Build_Windows_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Build_Windows_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Correctness_Rebuild|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Correctness_Rebuild|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[64144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=64144)|[PR 65008](https://github.com/dotnet/roslyn/pull/65008)|Build_Windows_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[62903](https://dev.azure.com/dnceng-public/public/_build/results?buildId=62903)|[PR 64957](https://github.com/dotnet/roslyn/pull/64957)|Test_Windows_Desktop_Debug_64|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_CoreClr_IOperation_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_64|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_Desktop_Debug_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_CoreClr_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56841)|Rolling|Test_Windows_CoreClr_Debug|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56316](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56316)|[PR 64803](https://github.com/dotnet/roslyn/pull/64803)|Test_Windows_Desktop_Release_32|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56316](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56316)|[PR 64803](https://github.com/dotnet/roslyn/pull/64803)|Test_Windows_CoreClr_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[56269](https://dev.azure.com/dnceng-public/public/_build/results?buildId=56269)|Rolling|Test_Windows_CoreClr_Debug_Single_Machine|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[51811](https://dev.azure.com/dnceng-public/public/_build/results?buildId=51811)|Rolling|Test_Windows_CoreClr_Release|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[39111](https://dev.azure.com/dnceng-public/public/_build/results?buildId=39111)|Rolling|Source-Build (Managed)|
|[roslyn-CI](https://dnceng-public.visualstudio.com/public/_build?definitionId=95)|[39111](https://dev.azure.com/dnceng-public/public/_build/results?buildId=39111)|Rolling|Build_Windows_Release|
Build Result Summary
|Day Hit Count|Week Hit Count|Month Hit Count|
|---|---|---|
|0|1|8|
<!-- runfo report end -->
| infrastructure | initializetoolset flakiness in correctness builds runfo tracking issue definition build kind job name managed build result summary day hit count week hit count month hit count | 1 |
222,521 | 17,080,823,632 | IssuesEvent | 2021-07-08 04:45:46 | razerr17/Compumundo_Hiper_Mega_Red_G4 | https://api.github.com/repos/razerr17/Compumundo_Hiper_Mega_Red_G4 | opened | Análisis Requerimientos #R3 y #R10 | documentation | Realizar los modelos SD y SR del requerimiento:
R3) Proporcionar a cada Docente Tutor el Avance Curricular de los estudiantes en riesgo académico que se encuentran bajo su tutoría. Esta información debe ser confidencial.
R10)c) Presentar semanalmente al Docente Tutor el reporte de asistencia de los estudiantes a su
cargo que se encuentran en riesgo académico. | 1.0 | Análisis Requerimientos #R3 y #R10 - Realizar los modelos SD y SR del requerimiento:
R3) Proporcionar a cada Docente Tutor el Avance Curricular de los estudiantes en riesgo académico que se encuentran bajo su tutoría. Esta información debe ser confidencial.
R10)c) Presentar semanalmente al Docente Tutor el reporte de asistencia de los estudiantes a su
cargo que se encuentran en riesgo académico. | non_infrastructure | análisis requerimientos y realizar los modelos sd y sr del requerimiento proporcionar a cada docente tutor el avance curricular de los estudiantes en riesgo académico que se encuentran bajo su tutoría esta información debe ser confidencial c presentar semanalmente al docente tutor el reporte de asistencia de los estudiantes a su cargo que se encuentran en riesgo académico | 0 |
213 | 2,578,593,392 | IssuesEvent | 2015-02-13 00:33:33 | FreeJournal/Free-Journal | https://api.github.com/repos/FreeJournal/Free-Journal | opened | Acquire and provision VPS's with required prerequisites | infrastructure | Acquire 4 VPS's in geographically diverse locations and provision | 1.0 | Acquire and provision VPS's with required prerequisites - Acquire 4 VPS's in geographically diverse locations and provision | infrastructure | acquire and provision vps s with required prerequisites acquire vps s in geographically diverse locations and provision | 1 |
829,593 | 31,883,763,298 | IssuesEvent | 2023-09-16 17:44:09 | Bentheburrito/cai | https://api.github.com/repos/Bentheburrito/cai | closed | Blurber sound bites for live sessions(?) | feature low priority bigboi | Could be cool to play sounds upon certain events like [Blurber](https://github.com/Bentheburrito/blurber) (could reuse much of the logic too probably) | 1.0 | Blurber sound bites for live sessions(?) - Could be cool to play sounds upon certain events like [Blurber](https://github.com/Bentheburrito/blurber) (could reuse much of the logic too probably) | non_infrastructure | blurber sound bites for live sessions could be cool to play sounds upon certain events like could reuse much of the logic too probably | 0 |
694,084 | 23,801,722,391 | IssuesEvent | 2022-09-03 12:00:22 | wso2/api-manager | https://api.github.com/repos/wso2/api-manager | opened | long query & blocking on mssql db | Type/Bug Priority/Normal | ### Description
Hi,
we noticed that we come across a situation where after certain scenarios, we get a long query report or blocking query on our MSSQL Server and the platform gets stuck till killing the DB query.
### Steps to Reproduce
1- Login to the Publisher Platform
2- Try to change a Policy on an API (specifically, we want to use XML to JSON & JSON to XML built-in policies)
3- Click on Save on the policy page.
then the panel gets stuck and we can't open and see other API details; after a couple of minutes, we get our long-running query errors and blocking queries
### Affected Component
APIM
### Version
4.1.0
### Environment Details (with versions)
1 Control Plane Node: Ubuntu Server 20.04 LTS
2 Gateway Nodes: Ubuntu Server 20.04 LTS
1 DB Node: SQL Server 2017
### Relevant Log Output
```shell
The Blocking query:
(@P0 int,@P1 varchar(8000))SELECT AUM.HTTP_METHOD, AUM.AUTH_SCHEME, AUM.URL_PATTERN, AUM.THROTTLING_TIER, AUM.MEDIATION_SCRIPT, ARSM.SCOPE_NAME, PROD_MAP.API_ID FROM AM_API_URL_MAPPING AUM LEFT JOIN AM_API_RESOURCE_SCOPE_MAPPING ARSM ON AUM.URL_MAPPING_ID = ARSM.URL_MAPPING_ID LEFT JOIN AM_API_PRODU
```
### Related Issues
_No response_
### Suggested Labels
_No response_ | 1.0 | long query & blocking on mssql db - ### Description
Hi,
we noticed that we come across a situation where after certain scenarios, we get a long query report or blocking query on our MSSQL Server and the platform gets stuck till killing the DB query.
### Steps to Reproduce
1- Login to the Publisher Platform
2- Try to change a Policy on an API (specifically, we want to use XML to JSON & JSON to XML built-in policies)
3- Click on Save on the policy page.
then the panel gets stuck and we can't open and see other API details; after a couple of minutes, we get our long-running query errors and blocking queries
### Affected Component
APIM
### Version
4.1.0
### Environment Details (with versions)
1 Control Plane Node: Ubuntu Server 20.04 LTS
2 Gateway Nodes: Ubuntu Server 20.04 LTS
1 DB Node: SQL Server 2017
### Relevant Log Output
```shell
The Blocking query:
(@P0 int,@P1 varchar(8000))SELECT AUM.HTTP_METHOD, AUM.AUTH_SCHEME, AUM.URL_PATTERN, AUM.THROTTLING_TIER, AUM.MEDIATION_SCRIPT, ARSM.SCOPE_NAME, PROD_MAP.API_ID FROM AM_API_URL_MAPPING AUM LEFT JOIN AM_API_RESOURCE_SCOPE_MAPPING ARSM ON AUM.URL_MAPPING_ID = ARSM.URL_MAPPING_ID LEFT JOIN AM_API_PRODU
```
### Related Issues
_No response_
### Suggested Labels
_No response_ | non_infrastructure | long query blocking on mssql db description hi we noticed that we come across a situation where after certain scenarios we get a long query report or blocking query on our mssql server and the platform gets stuck till killing the db query steps to reproduce login to the publisher platform try to change a policy on an api specifically we want to use xml to json json to xml built in policies click on save on the policy page then the panel gets stuck and we can t open and see other api details after a couple of minutes we get our long running query errors and blocking queries affected component apim version environment details with versions control plane node ubuntu server lts gateway nodes ubuntu server lts db node sql server relevant log output shell the blocking query int varchar select aum http method aum auth scheme aum url pattern aum throttling tier aum mediation script arsm scope name prod map api id from am api url mapping aum left join am api resource scope mapping arsm on aum url mapping id arsm url mapping id left join am api produ related issues no response suggested labels no response | 0 |
23,843 | 16,617,952,929 | IssuesEvent | 2021-06-02 19:22:29 | emory-libraries/blacklight-catalog | https://api.github.com/repos/emory-libraries/blacklight-catalog | closed | Create Blacklight SOLR restore script | Infrastructure | For purposes of disaster recovery, the Infrastructure Team is asked to create and document a script for restoring a blacklight solr instance from backup (ticket [537](https://app.zenhub.com/workspaces/blacklight-catalog-5f5f84a8a6d29939a0bc5d78/issues/emory-libraries/blacklight-catalog/537)).
- [x] Create script to restore SOLR blacklight collection
- Ideally one script should be usable across any environment
- [x] Document steps for collection restoration and verification
| 1.0 | Create Blacklight SOLR restore script - For purposes of disaster recovery, the Infrastructure Team is asked to create and document a script for restoring a blacklight solr instance from backup (ticket [537](https://app.zenhub.com/workspaces/blacklight-catalog-5f5f84a8a6d29939a0bc5d78/issues/emory-libraries/blacklight-catalog/537)).
- [x] Create script to restore SOLR blacklight collection
- Ideally one script should be usable across any environment
- [x] Document steps for collection restoration and verification
| infrastructure | create blacklight solr restore script for purposes of disaster recovery the infrastructure team is asked to create and document a script for restoring a blacklight solr instance from backup ticket create script to restore solr blacklight collection ideally one script should be usable across any environment document steps for collection restoration and verification | 1 |
32,776 | 26,988,809,582 | IssuesEvent | 2023-02-09 18:07:37 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | opened | Update test matrix for .net 8 | area-infrastructure | Update the OS's we test on for .Net 8 - Will involve updating the helix images defined in https://github.com/dotnet/aspnetcore/blob/main/eng/targets/Helix.Common.props | 1.0 | Update test matrix for .net 8 - Update the OS's we test on for .Net 8 - Will involve updating the helix images defined in https://github.com/dotnet/aspnetcore/blob/main/eng/targets/Helix.Common.props | infrastructure | update test matrix for net update the os s we test on for net will involve updating the helix images defined in | 1 |
148,020 | 19,526,984,963 | IssuesEvent | 2021-12-30 09:43:59 | SmartBear/soapui-raml-plugin | https://api.github.com/repos/SmartBear/soapui-raml-plugin | closed | CVE-2021-21342 (High) detected in xstream-1.4.12.jar - autoclosed | security vulnerability | ## CVE-2021-21342 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.12.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.12/xstream-1.4.12.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-3.5.0.jar (Root Library)
- ready-api-soapui-3.5.0.jar
- :x: **xstream-1.4.12.jar** (Vulnerable Library)
<p>Found in base branch: <b>next</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects. XStream creates therefore new instances based on these type information. An attacker can manipulate the processed input stream and replace or inject objects, that result in a server-side forgery request. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21342>CVE-2021-21342</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m">https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.12","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:3.5.0;com.smartbear:ready-api-soapui:3.5.0;com.thoughtworks.xstream:xstream:1.4.12","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16","isBinary":false}],"baseBranches":["next"],"vulnerabilityIdentifier":"CVE-2021-21342","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects. XStream creates therefore new instances based on these type information. An attacker can manipulate the processed input stream and replace or inject objects, that result in a server-side forgery request. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21342","cvss3Severity":"high","cvss3Score":"9.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-21342 (High) detected in xstream-1.4.12.jar - autoclosed - ## CVE-2021-21342 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.12.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.12/xstream-1.4.12.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-3.5.0.jar (Root Library)
- ready-api-soapui-3.5.0.jar
- :x: **xstream-1.4.12.jar** (Vulnerable Library)
<p>Found in base branch: <b>next</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects. XStream creates therefore new instances based on these type information. An attacker can manipulate the processed input stream and replace or inject objects, that result in a server-side forgery request. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21342>CVE-2021-21342</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m">https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.12","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:3.5.0;com.smartbear:ready-api-soapui:3.5.0;com.thoughtworks.xstream:xstream:1.4.12","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16","isBinary":false}],"baseBranches":["next"],"vulnerabilityIdentifier":"CVE-2021-21342","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects. XStream creates therefore new instances based on these type information. An attacker can manipulate the processed input stream and replace or inject objects, that result in a server-side forgery request. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21342","cvss3Severity":"high","cvss3Score":"9.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in xstream jar autoclosed cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy ready api soapui pro jar root library ready api soapui jar x xstream jar vulnerable library found in base branch next vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects xstream creates therefore new instances based on these type information an attacker can manipulate the processed input stream and replace or inject objects that result in a server side forgery request no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro com smartbear ready api soapui com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects xstream creates therefore new instances based on these type information an attacker can manipulate the processed input stream and replace or inject objects that result in a server side forgery request no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types if you rely on xstream default blacklist of the security framework you will have to use at least version vulnerabilityurl | 0 |
224 | 2,582,772,652 | IssuesEvent | 2015-02-15 16:56:50 | krazemon/repcoin | https://api.github.com/repos/krazemon/repcoin | closed | Investor should not have to add a category to invest in it | infrastructure | -When someone invests in a category, that category is added to their portfolio
-When someone revokes their last investment for a category, that category is removed from their portfolio
DEV
-Remove routes to add and delete an investor category
-Portfolio Table should no longer be editable
-Modal should no longer have prompt to be an investor
-Routes should no longer check if you are an investor before doing stuff
-Modal should no longer check if you are an investor to invest in someone
-Remove becoming an investor for a given category from the category page | 1.0 | Investor should not have to add a category to invest in it - -When someone invests in a category, that category is added to their portfolio
-When someone revokes their last investment for a category, that category is removed from their portfolio
DEV
-Remove routes to add and delete an investor category
-Portfolio Table should no longer be editable
-Modal should no longer have prompt to be an investor
-Routes should no longer check if you are an investor before doing stuff
-Modal should no longer check if you are an investor to invest in someone
-Remove becoming an investor for a given category from the category page | infrastructure | investor should not have to add a category to invest in it when someone invests in a category that category is added to their portfolio when someone revokes their last investment for a category that category is removed from their portfolio dev remove routes to add and delete an investor category portfolio table should no longer be editable modal should no longer have prompt to be an investor routes should no longer check if you are an investor before doing stuff modal should no longer check if you are an investor to invest in someone remove becoming an investor for a given category from the category page | 1 |
10,292 | 8,474,799,746 | IssuesEvent | 2018-10-24 17:07:51 | aiidateam/aiida_core | https://api.github.com/repos/aiidateam/aiida_core | closed | 2to3 as a pre-commit hook | topic/TestInfrastructure type/task | Add 2to3 as a (for now, optional) pre-commit linter to enforce some level of py3 compatibility for new / converted code. Begin with print statements only. | 1.0 | 2to3 as a pre-commit hook - Add 2to3 as a (for now, optional) pre-commit linter to enforce some level of py3 compatibility for new / converted code. Begin with print statements only. | infrastructure | as a pre commit hook add as a for now optional pre commit linter to enforce some level of compatibility for new converted code begin with print statements only | 1 |
32,375 | 26,657,138,840 | IssuesEvent | 2023-01-25 17:44:13 | practicalli/clojure | https://api.github.com/repos/practicalli/clojure | closed | Set up Clojure tools environment for using in GitHub Actions. | JVM infrastructure | Use a JDK docker image and add any specific version of Clojure to the image
- Clojure CLI tools
- Leiningen
- Boot
https://github.com/DeLaGuardo/setup-clojure
Example: install the Clojure tools on Ubuntu using DeLaGuardo/setup-clojure@2.0. You can see a workflow of one of my projects here: https://github.com/samply/blaze/blob/develop/.github/workflows/build.yml
| 1.0 | Set up Clojure tools environment for using in GitHub Actions. - Use a JDK docker image and add any specific version of Clojure to the image
- Clojure CLI tools
- Leiningen
- Boot
https://github.com/DeLaGuardo/setup-clojure
Example: install the Clojure tools on Ubuntu using DeLaGuardo/setup-clojure@2.0. You can see a workflow of one of my projects here: https://github.com/samply/blaze/blob/develop/.github/workflows/build.yml
| infrastructure | set up clojure tools environment for using in github actions use a jdk docker image and add any specific version of clojure to the image clojure cli tools leiningen boot example install the clojure tools on ubuntu using delaguardo setup clojure you can see a workflow of one of my projects here | 1 |
10,559 | 4,074,267,030 | IssuesEvent | 2016-05-28 09:54:46 | SleepyTrousers/EnderIO | https://api.github.com/repos/SleepyTrousers/EnderIO | closed | Soulbound not working with Baubles | Code Complete enhancement | The Soulbound enchantment is not working with items in the baubles slots.
Running EnderIO Version: 2.2.8.349
and Baubles Version: 1.0.1.10
on Minecraft 1.7.10
This happened after updating EnderIO from 2.2.7.325 to 2.2.8.349!
Help is much appreciated! | 1.0 | Soulbound not working with Baubles - The Soulbound enchantment is not working with items in the baubles slots.
Running EnderIO Version: 2.2.8.349
and Baubles Version: 1.0.1.10
on Minecraft 1.7.10
This happened after updating EnderIO from 2.2.7.325 to 2.2.8.349!
Help is much appreciated! | non_infrastructure | soulbound not working with baubles the soulbound enchantment is not working with items in the baubles slots running enderio version and baubles version on minecraft this happened after updating enderio from to help is much appreciated | 0 |
34,709 | 30,308,609,958 | IssuesEvent | 2023-07-10 11:15:07 | GSA-TTS/FAC | https://api.github.com/repos/GSA-TTS/FAC | closed | Get cypress specs running in CI | front end infrastructure | The FAC app uses Cypress to automate testing of the UI and end-to-end testing. The specs all run successfully against the app running locally, but currently fail when running in CI. At some point after the specs were ported over from the old FAC-Frontend app and added to the CI pipeline, some of the specs began failing intermittently and a bit unpredictably. Eventually most or possibly all of them began failing, despite no changes to the specs or the code under test.
Prior art: the implementation of this that worked all the way up until it didn't is in #862
## Acceptance Criteria
- [ ] All cypress specs run successfully in Github Actions | 1.0 | Get cypress specs running in CI - The FAC app uses Cypress to automate testing of the UI and end-to-end testing. The specs all run successfully against the app running locally, but currently fail when running in CI. At some point after the specs were ported over from the old FAC-Frontend app and added to the CI pipeline, some of the specs began failing intermittently and a bit unpredictably. Eventually most or possibly all of them began failing, despite no changes to the specs or the code under test.
Prior art: the implementation of this that worked all the way up until it didn't is in #862
## Acceptance Criteria
- [ ] All cypress specs run successfully in Github Actions | infrastructure | get cypress specs running in ci the fac app uses cypress to automate testing of the ui and end to end testing the specs all run successfully against the app running locally but currently fail when running in ci at some point after the specs were ported over from the old fac frontend app and added to the ci pipeline some of the specs began failing intermittently and a bit unpredictably eventually most or possibly all of them began failing despite no changes to the specs or the code under test prior art the implementation of this that worked all the way up until it didn t is in acceptance criteria all cypress specs run successfully in github actions | 1 |
11,292 | 9,084,200,528 | IssuesEvent | 2019-02-18 02:18:09 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Models visible to UI via Add model | interface/infrastructure question | It seems the old approach of manually adding models in models.cs has been removed and any model (implementing IModel) in the code is now visible in the UI via Add Model.
Can we add an Attribute (NoDisplay or such) that leaves a model hidden from the user (e.g. for base classes and models the user is not suppose to use directly)?
This attribute could be checked in Apsim.IsChildAllowable or Apsim.GetAllowableChildModels.
It might also be useful to add NotChildOf or InvalidParent (opposite of ValidParent) attribute to ensure some models do not appear under particular models in the tree. This would mean we can remove CLEM models from appearing under a Zone for non-CLEM users and simplify the add model window. | 1.0 | Models visible to UI via Add model - It seems the old approach of manually adding models in models.cs has been removed and any model (implementing IModel) in the code is now visible in the UI via Add Model.
Can we add an Attribute (NoDisplay or such) that leaves a model hidden from the user (e.g. for base classes and models the user is not suppose to use directly)?
This attribute could be checked in Apsim.IsChildAllowable or Apsim.GetAllowableChildModels.
It might also be useful to add NotChildOf or InvalidParent (opposite of ValidParent) attribute to ensure some models do not appear under particular models in the tree. This would mean we can remove CLEM models from appearing under a Zone for non-CLEM users and simplify the add model window. | infrastructure | models visible to ui via add model it seems the old approach of manually adding models in models cs has been removed and any model implementing imodel in the code is now visible in the ui via add model can we add an attribute nodisplay or such that leaves a model hidden from the user e g for base classes and models the user is not suppose to use directly this attribute could be checked in apsim ischildallowable or apsim getallowablechildmodels it might also be useful to add notchildof or invalidparent opposite of validparent attribute to ensure some models do not appear under particular models in the tree this would mean we can remove clem models from appearing under a zone for non clem users and simplify the add model window | 1 |
290,320 | 21,875,768,072 | IssuesEvent | 2022-05-19 09:59:14 | appsmithorg/appsmith | https://api.github.com/repos/appsmithorg/appsmith | closed | [Docs] #11937 [Bug]-[1500]:Upon adding a new JSObj/Query/Blank API's - EntityExplorer should show the new item in the expanded tree | Documentation User Education Pod | > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | New Developers Pod
**Parent Ticket** | #11937
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment | | 1.0 | [Docs] #11937 [Bug]-[1500]:Upon adding a new JSObj/Query/Blank API's - EntityExplorer should show the new item in the expanded tree - > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | New Developers Pod
**Parent Ticket** | #11937
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment | | non_infrastructure | upon adding a new jsobj query blank api s entityexplorer should show the new item in the expanded tree todo evaluate if this task is needed if not add the skip docs label on the parent ticket fill these fields prepare first draft add label ready for docs team field details pod new developers pod parent ticket engineer release date live date first draft auto assign priority environment | 0 |
19,845 | 3,265,300,131 | IssuesEvent | 2015-10-22 15:40:44 | giseislab/gismotools | https://api.github.com/repos/giseislab/gismotools | closed | get_antelope_traces fails when data files are missing | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
------------------------------------------------------------
scnl=scnlobject('SIT','BHZ','','');
ds = datasource('uaf_continuous');
startTime = '06-Jan-2000 10:41:42';
endTime = '06-Jan-2000 10:43:42';
w = waveform(ds,scnl,startTime,endTime);
What is the expected output? What do you see instead?
-----------------------------------------------------------
The above example errors out in an unintelligible way. I believe this is
because database entries actual exists for this scnl and time. However the
waveform files themselves have been taken offline. This failure should happen
frequently for waveform calls prior to October 2002. It is conceivable that it
would happen on rare occasions at other times. It is a problem that could occur
when calling any malformed .wfdisc table. (Ultimately this is the result of
poor database integrity. The fact is, though, it happens.)
I suspect the fix is a simple additional check in
load_antelope>get_antelope_traces. There is surely a formal database way to do
this check. I don't know it however. The question is how to ask the database
whether or not data claimed in a wfidisc row, actually exists on file?
Any thoughts on this last question would be welcomed.
```
Original issue reported on code.google.com by `mew...@alaska.edu` on 8 Nov 2011 at 9:41 | 1.0 | get_antelope_traces fails when data files are missing - ```
What steps will reproduce the problem?
------------------------------------------------------------
scnl=scnlobject('SIT','BHZ','','');
ds = datasource('uaf_continuous');
startTime = '06-Jan-2000 10:41:42';
endTime = '06-Jan-2000 10:43:42';
w = waveform(ds,scnl,startTime,endTime);
What is the expected output? What do you see instead?
-----------------------------------------------------------
The above example errors out in an unintelligible way. I believe this is
because database entries actual exists for this scnl and time. However the
waveform files themselves have been taken offline. This failure should happen
frequently for waveform calls prior to October 2002. It is conceivable that it
would happen on rare occasions at other times. It is a problem that could occur
when calling any malformed .wfdisc table. (Ultimately this is the result of
poor database integrity. The fact is, though, it happens.)
I suspect the fix is a simple additional check in
load_antelope>get_antelope_traces. There is surely a formal database way to do
this check. I don't know it however. The question is how to ask the database
whether or not data claimed in a wfidisc row, actually exists on file?
Any thoughts on this last question would be welcomed.
```
Original issue reported on code.google.com by `mew...@alaska.edu` on 8 Nov 2011 at 9:41 | non_infrastructure | get antelope traces fails when data files are missing what steps will reproduce the problem scnl scnlobject sit bhz ds datasource uaf continuous starttime jan endtime jan w waveform ds scnl starttime endtime what is the expected output what do you see instead the above example errors out in an unintelligible way i believe this is because database entries actual exists for this scnl and time however the waveform files themselves have been taken offline this failure should happen frequently for waveform calls prior to october it is conceivable that it would happen on rare occasions at other times it is a problem that could occur when calling any malformed wfdisc table ultimately this is the result of poor database integrity the fact is though it happens i suspect the fix is a simple additional check in load antelope get antelope traces there is surely a formal database way to do this check i don t know it however the question is how to ask the database whether or not data claimed in a wfidisc row actually exists on file any thoughts on this last question would be welcomed original issue reported on code google com by mew alaska edu on nov at | 0 |
31,750 | 26,076,891,809 | IssuesEvent | 2022-12-24 17:23:44 | iiif-prezi/iiif-prezi3 | https://api.github.com/repos/iiif-prezi/iiif-prezi3 | closed | Add explicit `griffe` requirement of >= 0.25.2 | infrastructure upstream | The issue discovered in https://github.com/iiif-prezi/iiif-prezi3/pull/145 and subsequently raised in https://github.com/mkdocstrings/griffe/issues/123 has been fixed thanks to fantastic work by @pawamoy, so we should enforce a minimum version in the dependencies (currently it's pulled in as a subdependency of mkdocstrings) to ensure consistent working docs builds across all environments. | 1.0 | Add explicit `griffe` requirement of >= 0.25.2 - The issue discovered in https://github.com/iiif-prezi/iiif-prezi3/pull/145 and subsequently raised in https://github.com/mkdocstrings/griffe/issues/123 has been fixed thanks to fantastic work by @pawamoy, so we should enforce a minimum version in the dependencies (currently it's pulled in as a subdependency of mkdocstrings) to ensure consistent working docs builds across all environments. | infrastructure | add explicit griffe requirement of the issue discovered in and subsequently raised in has been fixed thanks to fantastic work by pawamoy so we should enforce a minimum version in the dependencies currently it s pulled in as a subdependency of mkdocstrings to ensure consistent working docs builds across all environments | 1 |
16,335 | 11,938,879,139 | IssuesEvent | 2020-04-02 14:25:29 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Build System.Private.CoreLib in a separate job from the CoreCLR product build | area-Infrastructure-coreclr | This will allow us (in combination with #33128) to build the libraries product build in parallel with the CoreCLR product build, which will cut down official build times by 15-20 minutes and enable future improvements like #33067 which can cut down PR and CI times even further. | 1.0 | Build System.Private.CoreLib in a separate job from the CoreCLR product build - This will allow us (in combination with #33128) to build the libraries product build in parallel with the CoreCLR product build, which will cut down official build times by 15-20 minutes and enable future improvements like #33067 which can cut down PR and CI times even further. | infrastructure | build system private corelib in a separate job from the coreclr product build this will allow us in combination with to build the libraries product build in parallel with the coreclr product build which will cut down official build times by minutes and enable future improvements like which can cut down pr and ci times even further | 1 |
103,005 | 16,599,871,187 | IssuesEvent | 2021-06-01 17:51:08 | freedomofpress/securedrop | https://api.github.com/repos/freedomofpress/securedrop | closed | Update cryptography dependency | release blocker security | The `cryptography` library is currently pinned at version 3.2.1, because [3.3 dropped Python 3.5 support](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst#33---2020-12-08), which is the version of Python included with Ubuntu 16.04.
After the full removal of Xenial support (#5725), we should be in a good position to update the `cryptography` library as part of the SecureDrop 2.0.0 release. We can consider:
- [version 3.3.2](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst#332---2021-02-07), which as of this writing is the last release with security fixes, and which was released prior to the inclusion of Rust code
- the latest stable version at the time of release, which will introduce the Rust build requirement. | True | Update cryptography dependency - The `cryptography` library is currently pinned at version 3.2.1, because [3.3 dropped Python 3.5 support](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst#33---2020-12-08), which is the version of Python included with Ubuntu 16.04.
After the full removal of Xenial support (#5725), we should be in a good position to update the `cryptography` library as part of the SecureDrop 2.0.0 release. We can consider:
- [version 3.3.2](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst#332---2021-02-07), which as of this writing is the last release with security fixes, and which was released prior to the inclusion of Rust code
- the latest stable version at the time of release, which will introduce the Rust build requirement. | non_infrastructure | update cryptography dependency the cryptography library is currently pinned at version because which is the version of python included with ubuntu after the full removal of xenial support we should be in a good position to update the cryptography library as part of the securedrop release we can consider which as of this writing is the last release with security fixes and which was released prior to the inclusion of rust code the latest stable version at the time of release which will introduce the rust build requirement | 0 |
502,237 | 14,542,827,981 | IssuesEvent | 2020-12-15 16:09:40 | geosolutions-it/austrocontrol-C125 | https://api.github.com/repos/geosolutions-it/austrocontrol-C125 | closed | Client ID 17 - Storing and editing measurements | Accepted C125-2020-AUSTROCONTROL-Map2Imp Priority: High user feedback | Once a measurement is finished (e.g. by double clicking) or after reopening it from annotations it is not possible to select and edit it via the graphical view (only via adding coordinates in editor). Instead when clicking in the map a new mesurment is started. This is not visible for the user, because the dropdown menu always stays on the first element.
Expectation.: Dropdown always shows currently edited measurement + click on element in map enables to add new vertices via click in map. | 1.0 | Client ID 17 - Storing and editing measurements - Once a measurement is finished (e.g. by double clicking) or after reopening it from annotations it is not possible to select and edit it via the graphical view (only via adding coordinates in editor). Instead when clicking in the map a new mesurment is started. This is not visible for the user, because the dropdown menu always stays on the first element.
Expectation.: Dropdown always shows currently edited measurement + click on element in map enables to add new vertices via click in map. | non_infrastructure | client id storing and editing measurements once a measurement is finished e g by double clicking or after reopening it from annotations it is not possible to select and edit it via the graphical view only via adding coordinates in editor instead when clicking in the map a new mesurment is started this is not visible for the user because the dropdown menu always stays on the first element expectation dropdown always shows currently edited measurement click on element in map enables to add new vertices via click in map | 0 |
4,165 | 4,860,305,070 | IssuesEvent | 2016-11-14 01:28:26 | ContentMine/old_site | https://api.github.com/repos/ContentMine/old_site | closed | Email not sending from Discourse | bug infrastructure | Need to check SparkPost settings etc.
@skasberger I will take a look and may defer to you, but will try my best not to distract you from exams!
| 1.0 | Email not sending from Discourse - Need to check SparkPost settings etc.
@skasberger I will take a look and may defer to you, but will try my best not to distract you from exams!
| infrastructure | email not sending from discourse need to check sparkpost settings etc skasberger i will take a look and may defer to you but will try my best not to distract you from exams | 1 |
191,818 | 14,596,491,052 | IssuesEvent | 2020-12-20 16:04:07 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | hello-mr-code/terraform-oci: oci/autoscaling_auto_scaling_configuration_test.go; 14 LoC | fresh small test |
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/autoscaling_auto_scaling_configuration_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/autoscaling_auto_scaling_configuration_test.go#L508-L521)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to autoScalingConfigurationId is reassigned at line 512
[Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/autoscaling_auto_scaling_configuration_test.go#L508-L521)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for _, autoScalingConfigurationId := range autoScalingConfigurationIds {
if ok := SweeperDefaultResourceId[autoScalingConfigurationId]; !ok {
deleteAutoScalingConfigurationRequest := oci_auto_scaling.DeleteAutoScalingConfigurationRequest{}
deleteAutoScalingConfigurationRequest.AutoScalingConfigurationId = &autoScalingConfigurationId
deleteAutoScalingConfigurationRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "auto_scaling")
_, error := autoScalingClient.DeleteAutoScalingConfiguration(context.Background(), deleteAutoScalingConfigurationRequest)
if error != nil {
fmt.Printf("Error deleting AutoScalingConfiguration %s %s, It is possible that the resource is already deleted. Please verify manually \n", autoScalingConfigurationId, error)
continue
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
| 1.0 | hello-mr-code/terraform-oci: oci/autoscaling_auto_scaling_configuration_test.go; 14 LoC -
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/autoscaling_auto_scaling_configuration_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/autoscaling_auto_scaling_configuration_test.go#L508-L521)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to autoScalingConfigurationId is reassigned at line 512
[Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/autoscaling_auto_scaling_configuration_test.go#L508-L521)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for _, autoScalingConfigurationId := range autoScalingConfigurationIds {
if ok := SweeperDefaultResourceId[autoScalingConfigurationId]; !ok {
deleteAutoScalingConfigurationRequest := oci_auto_scaling.DeleteAutoScalingConfigurationRequest{}
deleteAutoScalingConfigurationRequest.AutoScalingConfigurationId = &autoScalingConfigurationId
deleteAutoScalingConfigurationRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "auto_scaling")
_, error := autoScalingClient.DeleteAutoScalingConfiguration(context.Background(), deleteAutoScalingConfigurationRequest)
if error != nil {
fmt.Printf("Error deleting AutoScalingConfiguration %s %s, It is possible that the resource is already deleted. Please verify manually \n", autoScalingConfigurationId, error)
continue
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
| non_infrastructure | hello mr code terraform oci oci autoscaling auto scaling configuration test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to autoscalingconfigurationid is reassigned at line click here to show the line s of go which triggered the analyzer go for autoscalingconfigurationid range autoscalingconfigurationids if ok sweeperdefaultresourceid ok deleteautoscalingconfigurationrequest oci auto scaling deleteautoscalingconfigurationrequest deleteautoscalingconfigurationrequest autoscalingconfigurationid autoscalingconfigurationid deleteautoscalingconfigurationrequest requestmetadata retrypolicy getretrypolicy true auto scaling error autoscalingclient deleteautoscalingconfiguration context background deleteautoscalingconfigurationrequest if error nil fmt printf error deleting autoscalingconfiguration s s it is possible that the resource is already deleted please verify manually n autoscalingconfigurationid error continue leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
104,460 | 8,972,660,808 | IssuesEvent | 2019-01-29 18:53:31 | DigitalCampus/django-oppia | https://api.github.com/repos/DigitalCampus/django-oppia | closed | SonarQube - complexity of viz/views.py summary_view function | enhancement medium priority refectoring/review testing | Refactor this function to reduce its Cognitive Complexity from 23 to the 15 allowed.
see: https://sonarcloud.io/project/issues?id=django_oppia&open=AWPfTOWDuoX1TGVkyu6u&resolved=false
Need to add some tests for this too | 1.0 | SonarQube - complexity of viz/views.py summary_view function - Refactor this function to reduce its Cognitive Complexity from 23 to the 15 allowed.
see: https://sonarcloud.io/project/issues?id=django_oppia&open=AWPfTOWDuoX1TGVkyu6u&resolved=false
Need to add some tests for this too | non_infrastructure | sonarqube complexity of viz views py summary view function refactor this function to reduce its cognitive complexity from to the allowed see need to add some tests for this too | 0 |
161,550 | 25,361,691,927 | IssuesEvent | 2022-11-20 23:46:06 | status-im/status-mobile | https://api.github.com/repos/status-im/status-mobile | closed | Implement Visual Test mechanism and configuration setup | feature new-design [quo2] | # Feature Issue
Reference this branch https://github.com/status-im/status-mobile/tree/jcaprani/visual-test
### User Story
As a developer I want to be able to create visual tests for the quo2 component library so that I can be confident I have not broke the ui on other components while refactoring implemented components.
### Description
Taking the work from this branch:
https://github.com/status-im/status-mobile/tree/jcaprani/visual-test
Clean up and implement the base structure and configuration for creating visual tests.
This includes adding some dependencies to the codebase -> Detox, jest-snapshot-testing.
### Acceptance Criteria
Visual test tooling installed in codebase.
A simple test can be run
Tool builds and runs without errors.
| 1.0 | Implement Visual Test mechanism and configuration setup - # Feature Issue
Reference this branch https://github.com/status-im/status-mobile/tree/jcaprani/visual-test
### User Story
As a developer I want to be able to create visual tests for the quo2 component library so that I can be confident I have not broke the ui on other components while refactoring implemented components.
### Description
Taking the work from this branch:
https://github.com/status-im/status-mobile/tree/jcaprani/visual-test
Clean up and implement the base structure and configuration for creating visual tests.
This includes adding some dependencies to the codebase -> Detox, jest-snapshot-testing.
### Acceptance Criteria
Visual test tooling installed in codebase.
A simple test can be run
Tool builds and runs without errors.
| non_infrastructure | implement visual test mechanism and configuration setup feature issue reference this branch user story as a developer i want to be able to create visual tests for the component library so that i can be confident i have not broke the ui on other components while refactoring implemented components description taking the work from this branch clean up and implement the base structure and configuration for creating visual tests this includes adding some dependencies to the codebase detox jest snapshot testing acceptance criteria visual test tooling installed in codebase a simple test can be run tool builds and runs without errors | 0 |
3,375 | 4,272,769,279 | IssuesEvent | 2016-07-13 15:24:07 | eslint/eslint | https://api.github.com/repos/eslint/eslint | closed | Create custom rule to verify a core rule has required `meta` properties | accepted infrastructure | This came out of a conversation with @pedrottimark, @ilyavolodin and @nzakas.
We would like to create a custom rule, internal to the eslint codebase, to make sure our core rules have all the required `meta` properties in the new rule format.
I'm willing to work on this. | 1.0 | Create custom rule to verify a core rule has required `meta` properties - This came out of a conversation with @pedrottimark, @ilyavolodin and @nzakas.
We would like to create a custom rule, internal to the eslint codebase, to make sure our core rules have all the required `meta` properties in the new rule format.
I'm willing to work on this. | infrastructure | create custom rule to verify a core rule has required meta properties this came out of a conversation with pedrottimark ilyavolodin and nzakas we would like to create a custom rule internal to the eslint codebase to make sure our core rules have all the required meta properties in the new rule format i m willing to work on this | 1 |
30,847 | 7,262,550,546 | IssuesEvent | 2018-02-19 06:49:03 | bramdejager/spcb | https://api.github.com/repos/bramdejager/spcb | closed | Error connecting to SharePoint Online | CodePlexMigrated Impact: Unassigned | There is an error when I try to connect to SharePoint Online:
For security reasons DTD is prohibited in this XML document. To enable DTD processing set the DtdProcessing property on XmlReaderSettings to Parse and pass the settings into XmlReader.Create method.
#### Attachments
[SPCB-UX31E-20160726.log](https://www.codeplex.com/Download/AttachmentDownload.ashx?ProjectName=spcb&WorkItemId=65089&FileAttachmentId=8763)
#### This work item was migrated from CodePlex
CodePlex work item ID: '65089'
Vote count: '2'
| 1.0 | Error connecting to SharePoint Online - There is an error when I try to connect to SharePoint Online:
For security reasons DTD is prohibited in this XML document. To enable DTD processing set the DtdProcessing property on XmlReaderSettings to Parse and pass the settings into XmlReader.Create method.
#### Attachments
[SPCB-UX31E-20160726.log](https://www.codeplex.com/Download/AttachmentDownload.ashx?ProjectName=spcb&WorkItemId=65089&FileAttachmentId=8763)
#### This work item was migrated from CodePlex
CodePlex work item ID: '65089'
Vote count: '2'
| non_infrastructure | error connecting to sharepoint online there is an error when i try to connect to sharepoint online for security reasons dtd is prohibited in this xml document to enable dtd processing set the dtdprocessing property on xmlreadersettings to parse and pass the settings into xmlreader create method attachments this work item was migrated from codeplex codeplex work item id vote count | 0 |
472 | 2,889,529,335 | IssuesEvent | 2015-06-13 14:17:33 | zerotier/ZeroTierOne | https://api.github.com/repos/zerotier/ZeroTierOne | closed | Use IP multicast with UDP instead of broadcast to perform local LAN peer discovery | compatibility enhancement | Improve on local network discovery by using well tested protocols based on IGMP, that are used by Apple TV, Skype, DLNA, and other great P2P software. Choose from Zeroconf, mDNS, UDP multicast, or Internet Connectivity Establishment (ICE). | True | Use IP multicast with UDP instead of broadcast to perform local LAN peer discovery - Improve on local network discovery by using well tested protocols based on IGMP, that are used by Apple TV, Skype, DLNA, and other great P2P software. Choose from Zeroconf, mDNS, UDP multicast, or Internet Connectivity Establishment (ICE). | non_infrastructure | use ip multicast with udp instead of broadcast to perform local lan peer discovery improve on local network discovery by using well tested protocols based on igmp that are used by apple tv skype dlna and other great software choose from zeroconf mdns udp multicast or internet connectivity establishment ice | 0 |
33,402 | 27,429,089,392 | IssuesEvent | 2023-03-01 23:02:38 | openxla/iree | https://api.github.com/repos/openxla/iree | opened | Add ability to measure peak memory usage in SHARK | infrastructure/shark | ### Request description
This includes memory usage for IREE, PyTorch and TF.
### What component(s) does this issue relate to?
_No response_
### Additional context
_No response_ | 1.0 | Add ability to measure peak memory usage in SHARK - ### Request description
This includes memory usage for IREE, PyTorch and TF.
### What component(s) does this issue relate to?
_No response_
### Additional context
_No response_ | infrastructure | add ability to measure peak memory usage in shark request description this includes memory usage for iree pytorch and tf what component s does this issue relate to no response additional context no response | 1 |
15,330 | 11,458,678,420 | IssuesEvent | 2020-02-07 04:23:25 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | Azure CLI Feature: Show OS Version, along with az --version | Core Infrastructure Output | To make it easier to add issues here in github, add the OS and other important info to the displayed output when we run `az --version`
Here is my current output
```
$ az --version
azure-cli (2.0.3)
acr (2.0.1)
acs (2.0.3)
appservice (0.1.3)
batch (2.0.1)
cloud (2.0.1)
component (2.0.1)
configure (2.0.3)
container (0.1.3)
core (2.0.3)
documentdb (0.1.3)
feedback (2.0.1)
find (0.0.2)
iot (0.1.3)
keyvault (2.0.1)
lab (0.0.2)
monitor (0.0.2)
network (2.0.3)
nspkg (2.0.0)
profile (2.0.3)
redis (0.2.0)
resource (2.0.3)
role (2.0.2)
sql (2.0.1)
storage (2.0.3)
vm (2.0.3)
Python (Darwin) 2.7.13 (default, Apr 4 2017, 08:47:57)
[GCC 4.2.1 Compatible Apple LLVM 8.1.0 (clang-802.0.38)]
``` | 1.0 | Azure CLI Feature: Show OS Version, along with az --version - To make it easier to add issues here in github, add the OS and other important info to the displayed output when we run `az --version`
Here is my current output
```
$ az --version
azure-cli (2.0.3)
acr (2.0.1)
acs (2.0.3)
appservice (0.1.3)
batch (2.0.1)
cloud (2.0.1)
component (2.0.1)
configure (2.0.3)
container (0.1.3)
core (2.0.3)
documentdb (0.1.3)
feedback (2.0.1)
find (0.0.2)
iot (0.1.3)
keyvault (2.0.1)
lab (0.0.2)
monitor (0.0.2)
network (2.0.3)
nspkg (2.0.0)
profile (2.0.3)
redis (0.2.0)
resource (2.0.3)
role (2.0.2)
sql (2.0.1)
storage (2.0.3)
vm (2.0.3)
Python (Darwin) 2.7.13 (default, Apr 4 2017, 08:47:57)
[GCC 4.2.1 Compatible Apple LLVM 8.1.0 (clang-802.0.38)]
``` | infrastructure | azure cli feature show os version along with az version to make it easier to add issues here in github add the os and other important info to the displayed output when we run az version here is my current output az version azure cli acr acs appservice batch cloud component configure container core documentdb feedback find iot keyvault lab monitor network nspkg profile redis resource role sql storage vm python darwin default apr | 1 |
14,488 | 10,885,143,194 | IssuesEvent | 2019-11-18 09:48:56 | tunapanda/swag | https://api.github.com/repos/tunapanda/swag | closed | Initialise CRUD Endpoints | backend points:8 priority:highest type:infrastructure | As a frontend engineer, I would like CRUD endpoints for the Ember app to communicate with.
- [x] Decide on all the endpoints needed, should more or less map to the database
- [x] Flesh out the basic models and routes using Koa.js and objection.js
Depends on #1 | 1.0 | Initialise CRUD Endpoints - As a frontend engineer, I would like CRUD endpoints for the Ember app to communicate with.
- [x] Decide on all the endpoints needed, should more or less map to the database
- [x] Flesh out the basic models and routes using Koa.js and objection.js
Depends on #1 | infrastructure | initialise crud endpoints as a frontend engineer i would like crud endpoints for the ember app to communicate with decide on all the endpoints needed should more or less map to the database flesh out the basic models and routes using koa js and objection js depends on | 1 |
328 | 2,648,956,000 | IssuesEvent | 2015-03-14 12:52:50 | jamming/sabina | https://api.github.com/repos/jamming/sabina | opened | Add optimize 'profile' | 0 - Backlog infrastructure | Create a profile to build the project with less debug information in classes and optimizing the bytecode.
<!---
@huboard:{"milestone_order":2,"order":0.25}
-->
| 1.0 | Add optimize 'profile' - Create a profile to build the project with less debug information in classes and optimizing the bytecode.
<!---
@huboard:{"milestone_order":2,"order":0.25}
-->
| infrastructure | add optimize profile create a profile to build the project with less debug information in classes and optimizing the bytecode huboard milestone order order | 1 |
57 | 2,506,380,545 | IssuesEvent | 2015-01-12 10:00:39 | RobDixonIII/Bloom | https://api.github.com/repos/RobDixonIII/Bloom | opened | Close Tab Buttons | feature infrastructure | Close a tabs in both the browser and the analytics application via a close button inside the tab. | 1.0 | Close Tab Buttons - Close a tabs in both the browser and the analytics application via a close button inside the tab. | infrastructure | close tab buttons close a tabs in both the browser and the analytics application via a close button inside the tab | 1 |
173,128 | 27,389,528,499 | IssuesEvent | 2023-02-28 15:25:15 | microsoft/pyright | https://api.github.com/repos/microsoft/pyright | closed | Regression with default match pattern for Never type on 1.1.196 | as designed | Take this simple example with `reportUnnecessaryComparison` set to `warning`
```python
from typing import assert_never
def get_foo() -> int | str:
...
foo = get_foo()
match foo:
case int():
...
case str():
...
case _: # Pattern will never be matched for subject type "Never" (reportUnnecessaryComparison)
assert_never(foo)
```
This was working fine on previous versions. | 1.0 | Regression with default match pattern for Never type on 1.1.196 - Take this simple example with `reportUnnecessaryComparison` set to `warning`
```python
from typing import assert_never
def get_foo() -> int | str:
...
foo = get_foo()
match foo:
case int():
...
case str():
...
case _: # Pattern will never be matched for subject type "Never" (reportUnnecessaryComparison)
assert_never(foo)
```
This was working fine on previous versions. | non_infrastructure | regression with default match pattern for never type on take this simple example with reportunnecessarycomparison set to warning python from typing import assert never def get foo int str foo get foo match foo case int case str case pattern will never be matched for subject type never reportunnecessarycomparison assert never foo this was working fine on previous versions | 0 |
30,239 | 24,681,429,233 | IssuesEvent | 2022-10-18 21:46:35 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | opened | E2E broken for user input in WP 4.x | P0 Type: Infrastructure | ## Bug Description
E2E jobs for WP 4.x started failing with the addition of the WP minimum version banner notice due to the user input tests expecting a user input banner notification instead.
Seeing as the minimum is about to be raised (and these jobs removed) and user input won't be release before then, these are safe to disable.
## Steps to reproduce
1. Review GHA logs for E2E jobs for WP 4.7 and WP 4.9
2. See consistent failures
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* The E2E tests for user input should be skipped for jobs that run with WP pre-5.2
* This is temporary, so we can be specific to jobs with WP 4.x
## Implementation Brief
* <!-- One or more bullet points for how to technically resolve the issue. Make sure to include changes to Storybook and visual regression tests where relevant. -->
### Test Coverage
* Test coverage is technically reduced here, although in an insignificant way as it will be removed in the near future anyways
## QA Brief
* <!-- One or more bullet points for how to test that the issue has been resolved. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| 1.0 | E2E broken for user input in WP 4.x - ## Bug Description
E2E jobs for WP 4.x started failing with the addition of the WP minimum version banner notice due to the user input tests expecting a user input banner notification instead.
Seeing as the minimum is about to be raised (and these jobs removed) and user input won't be release before then, these are safe to disable.
## Steps to reproduce
1. Review GHA logs for E2E jobs for WP 4.7 and WP 4.9
2. See consistent failures
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* The E2E tests for user input should be skipped for jobs that run with WP pre-5.2
* This is temporary, so we can be specific to jobs with WP 4.x
## Implementation Brief
* <!-- One or more bullet points for how to technically resolve the issue. Make sure to include changes to Storybook and visual regression tests where relevant. -->
### Test Coverage
* Test coverage is technically reduced here, although in an insignificant way as it will be removed in the near future anyways
## QA Brief
* <!-- One or more bullet points for how to test that the issue has been resolved. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| infrastructure | broken for user input in wp x bug description jobs for wp x started failing with the addition of the wp minimum version banner notice due to the user input tests expecting a user input banner notification instead seeing as the minimum is about to be raised and these jobs removed and user input won t be release before then these are safe to disable steps to reproduce review gha logs for jobs for wp and wp see consistent failures do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria the tests for user input should be skipped for jobs that run with wp pre this is temporary so we can be specific to jobs with wp x implementation brief test coverage test coverage is technically reduced here although in an insignificant way as it will be removed in the near future anyways qa brief changelog entry | 1 |
30,185 | 24,626,255,855 | IssuesEvent | 2022-10-16 15:02:42 | open-duelyst/duelyst | https://api.github.com/repos/open-duelyst/duelyst | opened | [P3] Implement pruning of ECR images | enhancement infrastructure | ## Summary
We can store about 58 versions of each image in ECR for free (50GB limit / 171MB per image * 5 images). We should implement some automated pruning of older images so we don't start to exceed the 50GB limit. | 1.0 | [P3] Implement pruning of ECR images - ## Summary
We can store about 58 versions of each image in ECR for free (50GB limit / 171MB per image * 5 images). We should implement some automated pruning of older images so we don't start to exceed the 50GB limit. | infrastructure | implement pruning of ecr images summary we can store about versions of each image in ecr for free limit per image images we should implement some automated pruning of older images so we don t start to exceed the limit | 1 |
243,702 | 26,287,394,774 | IssuesEvent | 2023-01-08 01:05:14 | Tim-Demo/JuiceShop | https://api.github.com/repos/Tim-Demo/JuiceShop | opened | CVE-2022-0536 (Medium) detected in follow-redirects-1.14.2.tgz | security vulnerability | ## CVE-2022-0536 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>follow-redirects-1.14.2.tgz</b></p></summary>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.2.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json,/frontend/node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- http-server-0.12.3.tgz (Root Library)
- http-proxy-1.18.1.tgz
- :x: **follow-redirects-1.14.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-Demo/JuiceShop/commit/ba236fd18ec3e6450d68d675bce1609d2e5d3230">ba236fd18ec3e6450d68d675bce1609d2e5d3230</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Exposure of Sensitive Information to an Unauthorized Actor in NPM follow-redirects prior to 1.14.8.
<p>Publish Date: 2022-02-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0536>CVE-2022-0536</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536</a></p>
<p>Release Date: 2022-02-09</p>
<p>Fix Resolution (follow-redirects): 1.14.8</p>
<p>Direct dependency fix Resolution (http-server): 0.13.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2022-0536 (Medium) detected in follow-redirects-1.14.2.tgz - ## CVE-2022-0536 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>follow-redirects-1.14.2.tgz</b></p></summary>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.2.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json,/frontend/node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- http-server-0.12.3.tgz (Root Library)
- http-proxy-1.18.1.tgz
- :x: **follow-redirects-1.14.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-Demo/JuiceShop/commit/ba236fd18ec3e6450d68d675bce1609d2e5d3230">ba236fd18ec3e6450d68d675bce1609d2e5d3230</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Exposure of Sensitive Information to an Unauthorized Actor in NPM follow-redirects prior to 1.14.8.
<p>Publish Date: 2022-02-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0536>CVE-2022-0536</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536</a></p>
<p>Release Date: 2022-02-09</p>
<p>Fix Resolution (follow-redirects): 1.14.8</p>
<p>Direct dependency fix Resolution (http-server): 0.13.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_infrastructure | cve medium detected in follow redirects tgz cve medium severity vulnerability vulnerable library follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules follow redirects package json frontend node modules follow redirects package json dependency hierarchy http server tgz root library http proxy tgz x follow redirects tgz vulnerable library found in head commit a href found in base branch main vulnerability details exposure of sensitive information to an unauthorized actor in npm follow redirects prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution follow redirects direct dependency fix resolution http server rescue worker helmet automatic remediation is available for this issue | 0 |
24,682 | 17,604,627,740 | IssuesEvent | 2021-08-17 15:33:42 | vmware-tanzu/tanzu-toolkit-for-visual-studio | https://api.github.com/repos/vmware-tanzu/tanzu-toolkit-for-visual-studio | opened | Move to H20 for our internal TAS platform | infrastructure | PCFOne has a limited time before EOL and our current PCF env needs some help. Need to look into H20 option. | 1.0 | Move to H20 for our internal TAS platform - PCFOne has a limited time before EOL and our current PCF env needs some help. Need to look into H20 option. | infrastructure | move to for our internal tas platform pcfone has a limited time before eol and our current pcf env needs some help need to look into option | 1 |
27,249 | 21,521,116,488 | IssuesEvent | 2022-04-28 14:17:31 | biocompute-objects/bco_api | https://api.github.com/repos/biocompute-objects/bco_api | opened | CI pipeline for Container build & Push | Infrastructure | There are a number of different way to deploy software. One common way is through containerization (Docker). To facilitate this, we should generate a CI/CD pipeline action to automatically generate and push a container to the registry when a new version is minted. | 1.0 | CI pipeline for Container build & Push - There are a number of different way to deploy software. One common way is through containerization (Docker). To facilitate this, we should generate a CI/CD pipeline action to automatically generate and push a container to the registry when a new version is minted. | infrastructure | ci pipeline for container build push there are a number of different way to deploy software one common way is through containerization docker to facilitate this we should generate a ci cd pipeline action to automatically generate and push a container to the registry when a new version is minted | 1 |
135,285 | 18,677,953,773 | IssuesEvent | 2021-10-31 21:50:55 | samq-ghdemo/easybuggy-private | https://api.github.com/repos/samq-ghdemo/easybuggy-private | opened | CVE-2016-4461 (High) detected in xwork-core-2.3.20.jar | security vulnerability | ## CVE-2016-4461 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xwork-core-2.3.20.jar</b></p></summary>
<p>Apache Struts 2</p>
<p>Library home page: <a href="http://struts.apache.org/">http://struts.apache.org/</a></p>
<p>Path to dependency file: easybuggy-private/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20/xwork-core-2.3.20.jar</p>
<p>
Dependency Hierarchy:
- vulnpackage-1.0.jar (Root Library)
- struts2-rest-plugin-2.3.20.jar
- struts2-core-2.3.20.jar
- :x: **xwork-core-2.3.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/easybuggy-private/commit/6ef2566cb8b39d29f6b8b76a1bd3860df7fac401">6ef2566cb8b39d29f6b8b76a1bd3860df7fac401</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Struts 2.x before 2.3.29 allows remote attackers to execute arbitrary code via a "%{}" sequence in a tag attribute, aka forced double OGNL evaluation. NOTE: this vulnerability exists because of an incomplete fix for CVE-2016-0785.
<p>Publish Date: 2017-10-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4461>CVE-2016-4461</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/struts/tree/STRUTS_2_3_29/">https://github.com/apache/struts/tree/STRUTS_2_3_29/</a></p>
<p>Release Date: 2017-10-16</p>
<p>Fix Resolution: org.apache.struts:struts2-core:2.3.29</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.struts.xwork","packageName":"xwork-core","packageVersion":"2.3.20","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"wss:vulnpackage:1.0;org.apache.struts:struts2-rest-plugin:2.3.20;org.apache.struts:struts2-core:2.3.20;org.apache.struts.xwork:xwork-core:2.3.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.struts:struts2-core:2.3.29"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-4461","vulnerabilityDetails":"Apache Struts 2.x before 2.3.29 allows remote attackers to execute arbitrary code via a \"%{}\" sequence in a tag attribute, aka forced double OGNL evaluation. NOTE: this vulnerability exists because of an incomplete fix for CVE-2016-0785.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4461","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2016-4461 (High) detected in xwork-core-2.3.20.jar - ## CVE-2016-4461 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xwork-core-2.3.20.jar</b></p></summary>
<p>Apache Struts 2</p>
<p>Library home page: <a href="http://struts.apache.org/">http://struts.apache.org/</a></p>
<p>Path to dependency file: easybuggy-private/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20/xwork-core-2.3.20.jar</p>
<p>
Dependency Hierarchy:
- vulnpackage-1.0.jar (Root Library)
- struts2-rest-plugin-2.3.20.jar
- struts2-core-2.3.20.jar
- :x: **xwork-core-2.3.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/easybuggy-private/commit/6ef2566cb8b39d29f6b8b76a1bd3860df7fac401">6ef2566cb8b39d29f6b8b76a1bd3860df7fac401</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Struts 2.x before 2.3.29 allows remote attackers to execute arbitrary code via a "%{}" sequence in a tag attribute, aka forced double OGNL evaluation. NOTE: this vulnerability exists because of an incomplete fix for CVE-2016-0785.
<p>Publish Date: 2017-10-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4461>CVE-2016-4461</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/struts/tree/STRUTS_2_3_29/">https://github.com/apache/struts/tree/STRUTS_2_3_29/</a></p>
<p>Release Date: 2017-10-16</p>
<p>Fix Resolution: org.apache.struts:struts2-core:2.3.29</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.struts.xwork","packageName":"xwork-core","packageVersion":"2.3.20","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"wss:vulnpackage:1.0;org.apache.struts:struts2-rest-plugin:2.3.20;org.apache.struts:struts2-core:2.3.20;org.apache.struts.xwork:xwork-core:2.3.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.struts:struts2-core:2.3.29"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-4461","vulnerabilityDetails":"Apache Struts 2.x before 2.3.29 allows remote attackers to execute arbitrary code via a \"%{}\" sequence in a tag attribute, aka forced double OGNL evaluation. NOTE: this vulnerability exists because of an incomplete fix for CVE-2016-0785.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4461","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in xwork core jar cve high severity vulnerability vulnerable library xwork core jar apache struts library home page a href path to dependency file easybuggy private pom xml path to vulnerable library home wss scanner repository org apache struts xwork xwork core xwork core jar dependency hierarchy vulnpackage jar root library rest plugin jar core jar x xwork core jar vulnerable library found in head commit a href found in base branch master vulnerability details apache struts x before allows remote attackers to execute arbitrary code via a sequence in a tag attribute aka forced double ognl evaluation note this vulnerability exists because of an incomplete fix for cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache struts core isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree wss vulnpackage org apache struts rest plugin org apache struts core org apache struts xwork xwork core isminimumfixversionavailable true minimumfixversion org apache struts core basebranches vulnerabilityidentifier cve vulnerabilitydetails apache struts x before allows remote attackers to execute arbitrary code via a sequence in a tag attribute aka forced double ognl evaluation note this vulnerability exists because of an incomplete fix for cve vulnerabilityurl | 0 |
9,616 | 8,053,640,727 | IssuesEvent | 2018-08-02 00:12:44 | HearthSim/hsredshift | https://api.github.com/repos/HearthSim/hsredshift | closed | Put hsredshift UDFs on a diet | cleanup infrastructure story | As an operational stakeholder in the hsredshift reporting framework, I need to ensure that the Python UDFs executed by the framework stay well within the resource limits established by AWS, so that my ad hoc queries do not fail.
Several of our UDFs have apparently become too large to execute on the cluster. It's a bit unclear why this is, since the volume of data occupied by card metadata is only ~15MB even by the most generous estimate. AWS does not document what the memory threshold for Python UDFs is, nor precisely what the lifecycle of Python UDFs is on the cluster. But we believe that we've been well under the limit in the past, and there are some concise steps we can take to reduce our memory footprint again. So we should switch - where possible - to SQL implementations joining against the `card` table, and we can also remove the UDFs we no longer use.
This story is complete when:
- [x] The `f_card_info` UDF (and all other unused UDFs) has been removed from the hsredshift codebase
- [x] All unused UDFs have been dropped from the production database
- [x] The `DBF_TO_CARD` dict has been modified to store only card name and cost and only include _collectible_ cards.
- [ ] The `f_card_url` UDF has been replaced with a SQL-based implementation
| 1.0 | Put hsredshift UDFs on a diet - As an operational stakeholder in the hsredshift reporting framework, I need to ensure that the Python UDFs executed by the framework stay well within the resource limits established by AWS, so that my ad hoc queries do not fail.
Several of our UDFs have apparently become too large to execute on the cluster. It's a bit unclear why this is, since the volume of data occupied by card metadata is only ~15MB even by the most generous estimate. AWS does not document what the memory threshold for Python UDFs is, nor precisely what the lifecycle of Python UDFs is on the cluster. But we believe that we've been well under the limit in the past, and there are some concise steps we can take to reduce our memory footprint again. So we should switch - where possible - to SQL implementations joining against the `card` table, and we can also remove the UDFs we no longer use.
This story is complete when:
- [x] The `f_card_info` UDF (and all other unused UDFs) has been removed from the hsredshift codebase
- [x] All unused UDFs have been dropped from the production database
- [x] The `DBF_TO_CARD` dict has been modified to store only card name and cost and only include _collectible_ cards.
- [ ] The `f_card_url` UDF has been replaced with a SQL-based implementation
| infrastructure | put hsredshift udfs on a diet as an operational stakeholder in the hsredshift reporting framework i need to ensure that the python udfs executed by the framework stay well within the resource limits established by aws so that my ad hoc queries do not fail several of our udfs have apparently become too large to execute on the cluster it s a bit unclear why this is since the volume of data occupied by card metadata is only even by the most generous estimate aws does not document what the memory threshold for python udfs is nor precisely what the lifecycle of python udfs is on the cluster but we believe that we ve been well under the limit in the past and there are some concise steps we can take to reduce our memory footprint again so we should switch where possible to sql implementations joining against the card table and we can also remove the udfs we no longer use this story is complete when the f card info udf and all other unused udfs has been removed from the hsredshift codebase all unused udfs have been dropped from the production database the dbf to card dict has been modified to store only card name and cost and only include collectible cards the f card url udf has been replaced with a sql based implementation | 1 |
540,594 | 15,814,201,288 | IssuesEvent | 2021-04-05 09:03:52 | threefoldfoundation/home | https://api.github.com/repos/threefoldfoundation/home | opened | WEB - web_2.8 release | priority_major | **DESCRIPTION:**
There are urgent updates across our websites needed before 8th April when we release Casper announcement.
**DEADLINE:** 8th April
**TO DO**:
- [ ] Show TFT on Interstellar: https://github.com/threefoldfoundation/home/issues/157
- [ ] Partners needs to be spotless: https://github.com/threefoldfoundation/data_threefold/issues/31
- [ ] Update TFT listings: https://github.com/threefoldfoundation/home/issues/160
- [ ] Footer added for DT: https://github.com/threefoldfoundation/www_threefold_twin/issues/57
- [ ] Add Casper on homepage as partner: https://github.com/threefoldfoundation/www_threefold_io/issues/42
- [ ] Prepare Casper page: https://github.com/threefoldfoundation/data_threefold/issues/27
- [ ] Casper announcement: https://github.com/threefoldfoundation/home_private/issues/19
- [ ] Add Casper app to marketplace: https://github.com/threefoldfoundation/www_threefold_marketplace/issues/33, https://github.com/threefoldfoundation/www_threefold_marketplace/issues/34 | 1.0 | WEB - web_2.8 release - **DESCRIPTION:**
There are urgent updates across our websites needed before 8th April when we release Casper announcement.
**DEADLINE:** 8th April
**TO DO**:
- [ ] Show TFT on Interstellar: https://github.com/threefoldfoundation/home/issues/157
- [ ] Partners needs to be spotless: https://github.com/threefoldfoundation/data_threefold/issues/31
- [ ] Update TFT listings: https://github.com/threefoldfoundation/home/issues/160
- [ ] Footer added for DT: https://github.com/threefoldfoundation/www_threefold_twin/issues/57
- [ ] Add Casper on homepage as partner: https://github.com/threefoldfoundation/www_threefold_io/issues/42
- [ ] Prepare Casper page: https://github.com/threefoldfoundation/data_threefold/issues/27
- [ ] Casper announcement: https://github.com/threefoldfoundation/home_private/issues/19
- [ ] Add Casper app to marketplace: https://github.com/threefoldfoundation/www_threefold_marketplace/issues/33, https://github.com/threefoldfoundation/www_threefold_marketplace/issues/34 | non_infrastructure | web web release description there are urgent updates across our websites needed before april when we release casper announcement deadline april to do show tft on interstellar partners needs to be spotless update tft listings footer added for dt add casper on homepage as partner prepare casper page casper announcement add casper app to marketplace | 0 |
256,784 | 27,561,721,689 | IssuesEvent | 2023-03-07 22:42:12 | samqws-marketing/amzn-ion-hive-serde | https://api.github.com/repos/samqws-marketing/amzn-ion-hive-serde | closed | CVE-2020-9548 (High) detected in jackson-databind-2.6.5.jar - autoclosed | Mend: dependency security vulnerability | ## CVE-2020-9548 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /integration-test/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- hive-serde-2.3.9.jar (Root Library)
- hive-common-2.3.9.jar
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/amzn-ion-hive-serde/commit/ffb6641ebb10aac58bb7eec412635e91e79fac24">ffb6641ebb10aac58bb7eec412635e91e79fac24</a></p>
<p>Found in base branch: <b>0.3.0</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core).
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9548>CVE-2020-9548</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548</a></p>
<p>Release Date: 2020-03-02</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.4</p>
<p>Direct dependency fix Resolution (org.apache.hive:hive-serde): 3.0.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2020-9548 (High) detected in jackson-databind-2.6.5.jar - autoclosed - ## CVE-2020-9548 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /integration-test/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- hive-serde-2.3.9.jar (Root Library)
- hive-common-2.3.9.jar
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/amzn-ion-hive-serde/commit/ffb6641ebb10aac58bb7eec412635e91e79fac24">ffb6641ebb10aac58bb7eec412635e91e79fac24</a></p>
<p>Found in base branch: <b>0.3.0</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core).
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9548>CVE-2020-9548</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548</a></p>
<p>Release Date: 2020-03-02</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.4</p>
<p>Direct dependency fix Resolution (org.apache.hive:hive-serde): 3.0.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_infrastructure | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file integration test build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy hive serde jar root library hive common jar x jackson databind jar vulnerable library found in head commit a href found in base branch vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpconfig aka anteros core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org apache hive hive serde check this box to open an automated fix pr | 0 |
26,730 | 20,616,232,087 | IssuesEvent | 2022-03-07 13:30:46 | LibraryOfCongress/concordia | https://api.github.com/repos/LibraryOfCongress/concordia | opened | Can Github Actions replace our current use of Jenkins on EC2 | infrastructure operations | Can Github Actions replace our current use of Jenkins on EC2? | 1.0 | Can Github Actions replace our current use of Jenkins on EC2 - Can Github Actions replace our current use of Jenkins on EC2? | infrastructure | can github actions replace our current use of jenkins on can github actions replace our current use of jenkins on | 1 |
21,163 | 28,136,771,621 | IssuesEvent | 2023-04-01 13:22:59 | firebase/firebase-cpp-sdk | https://api.github.com/repos/firebase/firebase-cpp-sdk | reopened | [C++] Nightly Integration Testing Report for Firestore | type: process nightly-testing |
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 73ce6feb70d3e830676aafa1d0ded64a57f07fb8
Last updated: Sat Apr 1 04:48 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4582570173)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 73ce6feb70d3e830676aafa1d0ded64a57f07fb8
Last updated: Fri Mar 31 09:55 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4574922030)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit 73ce6feb70d3e830676aafa1d0ded64a57f07fb8
Last updated: Sat Apr 1 04:39 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4582793169)**
| 1.0 | [C++] Nightly Integration Testing Report for Firestore -
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 73ce6feb70d3e830676aafa1d0ded64a57f07fb8
Last updated: Sat Apr 1 04:48 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4582570173)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 73ce6feb70d3e830676aafa1d0ded64a57f07fb8
Last updated: Fri Mar 31 09:55 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4574922030)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit 73ce6feb70d3e830676aafa1d0ded64a57f07fb8
Last updated: Sat Apr 1 04:39 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4582793169)**
| non_infrastructure | nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated sat apr pdt ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated fri mar pdt ✅ nbsp integration test succeeded requested by on commit last updated sat apr pdt | 0 |
79,576 | 28,432,439,113 | IssuesEvent | 2023-04-15 00:20:55 | zealdocs/zeal | https://api.github.com/repos/zealdocs/zeal | closed | Pandas docset displays in a tiny frame | type/defect resolution/awaiting-response scope/ui/webview | Zeal 0.6.1 (Xubuntu 20.04.5 LTS)
Pandas docset "Date: Sep 19, 2022 Version: 1.5.0"
The Pandas docset displays in a vary narrow (2 lines?) frame, inside the main content pane in Zeal. See attached screenshot.
Other docsets display correctly.

| 1.0 | Pandas docset displays in a tiny frame - Zeal 0.6.1 (Xubuntu 20.04.5 LTS)
Pandas docset "Date: Sep 19, 2022 Version: 1.5.0"
The Pandas docset displays in a vary narrow (2 lines?) frame, inside the main content pane in Zeal. See attached screenshot.
Other docsets display correctly.

| non_infrastructure | pandas docset displays in a tiny frame zeal xubuntu lts pandas docset date sep version the pandas docset displays in a vary narrow lines frame inside the main content pane in zeal see attached screenshot other docsets display correctly | 0 |
19,212 | 13,207,401,078 | IssuesEvent | 2020-08-14 22:57:51 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | git (Trac #89) | Incomplete Migration Migrated from Trac enhancement infrastructure | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/89">https://code.icecube.wisc.edu/projects/icecube/ticket/89</a>, reported by troyand owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-08-29T01:05:40",
"_ts": "1188349540000000",
"description": "git git git git git especially on pa-pub, you can't git through there at all atm",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"time": "2007-08-09T22:01:12",
"component": "infrastructure",
"summary": "git",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cgils",
"type": "enhancement"
}
```
</p>
</details>
| 1.0 | git (Trac #89) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/89">https://code.icecube.wisc.edu/projects/icecube/ticket/89</a>, reported by troyand owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-08-29T01:05:40",
"_ts": "1188349540000000",
"description": "git git git git git especially on pa-pub, you can't git through there at all atm",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"time": "2007-08-09T22:01:12",
"component": "infrastructure",
"summary": "git",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cgils",
"type": "enhancement"
}
```
</p>
</details>
| infrastructure | git trac migrated from json status closed changetime ts description git git git git git especially on pa pub you can t git through there at all atm reporter troy cc resolution fixed time component infrastructure summary git priority normal keywords milestone owner cgils type enhancement | 1 |
2,403 | 3,667,422,606 | IssuesEvent | 2016-02-20 00:46:47 | Microsoft/TypeScript | https://api.github.com/repos/Microsoft/TypeScript | closed | Questions about your testing strategy | Discussion Infrastructure Question | Your testing strategy is awesome and I'm very impressed. It is one of the big golden eggs in this project. I just wonder, if it where you guys who invented it? Is there a name for it and if there is more info about it? | 1.0 | Questions about your testing strategy - Your testing strategy is awesome and I'm very impressed. It is one of the big golden eggs in this project. I just wonder, if it where you guys who invented it? Is there a name for it and if there is more info about it? | infrastructure | questions about your testing strategy your testing strategy is awesome and i m very impressed it is one of the big golden eggs in this project i just wonder if it where you guys who invented it is there a name for it and if there is more info about it | 1 |
381,790 | 11,287,938,229 | IssuesEvent | 2020-01-16 06:25:06 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | closed | Unable to use 2 or more deploy tools in 1 skaffold file | area/deploy deploy/helm deploy/kubectl deploy/kustomize kind/feature-request priority/p2 | Following [k8s docs](https://kubernetes.io/docs/concepts/configuration/secret/#creating-a-secret-from-generator) want to create secrets with `kustomize` and use `kubectl` for other configs.
### Expected behavior
No error
### Actual behavior
```bash
$ skaffold diagnose
FATA[0000] creating runner: invalid skaffold config: only one element in set deploy can be set. got HelmDeploy and KubectlDeploy
```
### Information
- Skaffold version: v0.38.0
- Operating system: Mac Os 10.14.6
- Contents of skaffold.yaml:
```yaml
apiVersion: skaffold/v1beta14
kind: Config
build:
tagPolicy:
envTemplate:
template: "{{.RELEASE}}-{{.IMAGE_NAME}}"
artifacts:
- image: core
context: core
docker:
dockerfile: docker/development/Dockerfile
- image: pubsub_emulator
context: ../pubsub_emulator
docker:
dockerfile: ../pubsub_emulator/Dockerfile
deploy:
kubectl:
manifests:
- support/rbac/endpoints.yaml
- support/pubsub_emulator/k8s.yaml
kustomize:
path: support/env_vars
helm:
releases:
- name: core
chartPath: core/charts/poker-core
valuesFiles:
- core/charts/core/values.yaml
- core/charts/core/values.development.yaml
```
| 1.0 | Unable to use 2 or more deploy tools in 1 skaffold file - Following [k8s docs](https://kubernetes.io/docs/concepts/configuration/secret/#creating-a-secret-from-generator) want to create secrets with `kustomize` and use `kubectl` for other configs.
### Expected behavior
No error
### Actual behavior
```bash
$ skaffold diagnose
FATA[0000] creating runner: invalid skaffold config: only one element in set deploy can be set. got HelmDeploy and KubectlDeploy
```
### Information
- Skaffold version: v0.38.0
- Operating system: Mac Os 10.14.6
- Contents of skaffold.yaml:
```yaml
apiVersion: skaffold/v1beta14
kind: Config
build:
tagPolicy:
envTemplate:
template: "{{.RELEASE}}-{{.IMAGE_NAME}}"
artifacts:
- image: core
context: core
docker:
dockerfile: docker/development/Dockerfile
- image: pubsub_emulator
context: ../pubsub_emulator
docker:
dockerfile: ../pubsub_emulator/Dockerfile
deploy:
kubectl:
manifests:
- support/rbac/endpoints.yaml
- support/pubsub_emulator/k8s.yaml
kustomize:
path: support/env_vars
helm:
releases:
- name: core
chartPath: core/charts/poker-core
valuesFiles:
- core/charts/core/values.yaml
- core/charts/core/values.development.yaml
```
| non_infrastructure | unable to use or more deploy tools in skaffold file following want to create secrets with kustomize and use kubectl for other configs expected behavior no error actual behavior bash skaffold diagnose fata creating runner invalid skaffold config only one element in set deploy can be set got helmdeploy and kubectldeploy information skaffold version operating system mac os contents of skaffold yaml yaml apiversion skaffold kind config build tagpolicy envtemplate template release image name artifacts image core context core docker dockerfile docker development dockerfile image pubsub emulator context pubsub emulator docker dockerfile pubsub emulator dockerfile deploy kubectl manifests support rbac endpoints yaml support pubsub emulator yaml kustomize path support env vars helm releases name core chartpath core charts poker core valuesfiles core charts core values yaml core charts core values development yaml | 0 |
307,001 | 26,511,145,665 | IssuesEvent | 2023-01-18 17:11:08 | gravitational/teleport | https://api.github.com/repos/gravitational/teleport | opened | `tsh db connect` fails when using `mariadb` and the proxy is in separate port mode. | bug testplan database-access db/mysql | Expected behavior:
`tsh db connect` should succeed
Current behavior:
```
ERROR 2002 (HY000): Can't connect to local server through socket '/tmp/mysql.sock' (2)
```
Bug details:
- Teleport version: v12 testplan
- Recreation steps:
- Proxy in "separate" mode
- Debug logs
```
$ tsh db connect -d --cluster $awsdev --db-user alice steve-rds
...
[TSH] DEBU /opt/homebrew/opt/mariadb/bin/mariadb --user alice --ssl-key /Users/stevehuang/.tsh_root/keys/teleport.root.dev.aws.stevexin.me/alb --ssl-ca /Users/stevehuang/.tsh_root/keys/teleport.root.dev.aws.stevexin.me/cas/teleport.root.dev.aws.stevexin.me.pem --ssl-cert /Users/stevehuang/.tsh_root/keys/teleport.root.dev.aws.stevexin.me/alb-db/teleport.dev.aws.stevexin.me/steve-rds-x509.pem --ssl-verify-server-cert tsh/db.go:759
ERROR 2002 (HY000): Can't connect to local server through socket '/tmp/mysql.sock' (2)
ERROR REPORT:
Original Error: *exec.ExitError exit status 1
Stack Trace:
github.com/gravitational/teleport/lib/client/db/dbcmd/error.go:56 github.com/gravitational/teleport/lib/client/db/dbcmd.ConvertCommandError
github.com/gravitational/teleport/tool/tsh/db.go:772 main.onDatabaseConnect
github.com/gravitational/teleport/tool/tsh/tsh.go:1128 main.Run
github.com/gravitational/teleport/tool/tsh/tsh.go:472 main.main
```
Note that there is no port or host information in the composed `mariadb` command.
Seems this is not a regression (never worked before). `tsh db connect` works fine in "multiplex" mode or when using the "mysql" client. | 1.0 | `tsh db connect` fails when using `mariadb` and the proxy is in separate port mode. - Expected behavior:
`tsh db connect` should succeed
Current behavior:
```
ERROR 2002 (HY000): Can't connect to local server through socket '/tmp/mysql.sock' (2)
```
Bug details:
- Teleport version: v12 testplan
- Recreation steps:
- Proxy in "separate" mode
- Debug logs
```
$ tsh db connect -d --cluster $awsdev --db-user alice steve-rds
...
[TSH] DEBU /opt/homebrew/opt/mariadb/bin/mariadb --user alice --ssl-key /Users/stevehuang/.tsh_root/keys/teleport.root.dev.aws.stevexin.me/alb --ssl-ca /Users/stevehuang/.tsh_root/keys/teleport.root.dev.aws.stevexin.me/cas/teleport.root.dev.aws.stevexin.me.pem --ssl-cert /Users/stevehuang/.tsh_root/keys/teleport.root.dev.aws.stevexin.me/alb-db/teleport.dev.aws.stevexin.me/steve-rds-x509.pem --ssl-verify-server-cert tsh/db.go:759
ERROR 2002 (HY000): Can't connect to local server through socket '/tmp/mysql.sock' (2)
ERROR REPORT:
Original Error: *exec.ExitError exit status 1
Stack Trace:
github.com/gravitational/teleport/lib/client/db/dbcmd/error.go:56 github.com/gravitational/teleport/lib/client/db/dbcmd.ConvertCommandError
github.com/gravitational/teleport/tool/tsh/db.go:772 main.onDatabaseConnect
github.com/gravitational/teleport/tool/tsh/tsh.go:1128 main.Run
github.com/gravitational/teleport/tool/tsh/tsh.go:472 main.main
```
Note that there is no port or host information in the composed `mariadb` command.
Seems this is not a regression (never worked before). `tsh db connect` works fine in "multiplex" mode or when using the "mysql" client. | non_infrastructure | tsh db connect fails when using mariadb and the proxy is in separate port mode expected behavior tsh db connect should succeed current behavior error can t connect to local server through socket tmp mysql sock bug details teleport version testplan recreation steps proxy in separate mode debug logs tsh db connect d cluster awsdev db user alice steve rds debu opt homebrew opt mariadb bin mariadb user alice ssl key users stevehuang tsh root keys teleport root dev aws stevexin me alb ssl ca users stevehuang tsh root keys teleport root dev aws stevexin me cas teleport root dev aws stevexin me pem ssl cert users stevehuang tsh root keys teleport root dev aws stevexin me alb db teleport dev aws stevexin me steve rds pem ssl verify server cert tsh db go error can t connect to local server through socket tmp mysql sock error report original error exec exiterror exit status stack trace github com gravitational teleport lib client db dbcmd error go github com gravitational teleport lib client db dbcmd convertcommanderror github com gravitational teleport tool tsh db go main ondatabaseconnect github com gravitational teleport tool tsh tsh go main run github com gravitational teleport tool tsh tsh go main main note that there is no port or host information in the composed mariadb command seems this is not a regression never worked before tsh db connect works fine in multiplex mode or when using the mysql client | 0 |
16,488 | 4,055,334,902 | IssuesEvent | 2016-05-24 15:11:05 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | Warnings while building Sphinx documentation | in progress priority: low team: software core type: documentation | ```
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/faq.rst:5: ERROR: Document or section may not begin with a transition.
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/sublime_text.rst:44: WARNING: Inline emphasis start-string without end-string.
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/reviewable.rst:: WARNING: document isn't included in any toctree
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/unit_testing_instructions.rst:: WARNING: document isn't included in any toctree
``` | 1.0 | Warnings while building Sphinx documentation - ```
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/faq.rst:5: ERROR: Document or section may not begin with a transition.
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/sublime_text.rst:44: WARNING: Inline emphasis start-string without end-string.
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/reviewable.rst:: WARNING: document isn't included in any toctree
/home/ubuntu/workspace/experimental/06107e1d/drake/doc/unit_testing_instructions.rst:: WARNING: document isn't included in any toctree
``` | non_infrastructure | warnings while building sphinx documentation home ubuntu workspace experimental drake doc faq rst error document or section may not begin with a transition home ubuntu workspace experimental drake doc sublime text rst warning inline emphasis start string without end string home ubuntu workspace experimental drake doc reviewable rst warning document isn t included in any toctree home ubuntu workspace experimental drake doc unit testing instructions rst warning document isn t included in any toctree | 0 |
14,524 | 10,919,079,861 | IssuesEvent | 2019-11-21 18:13:46 | patternfly/patternfly-org | https://api.github.com/repos/patternfly/patternfly-org | closed | Add ability to reach older version documentation for PF4 | PF4 DEV infrastructure | Documentation released that is out of date needs to be publicly accessible - this can be reached using a version switcher, archived somewhere (or in some other way). Users should be able to quickly identify which version the documentation refers to and view documentation for older releases.
Discussion is needed about whether we do this for major breaking changes (so when @patternfly/patternfly bumps from 2.x to 3.x - do we archive and allow people to switch back to 2.x) and then just update the 3.x visible documentation until we hit 4.x. We also need to discuss what this means for the @patternfly/react-* packages which can move independently. A version switcher for those packages could be harder to embed. We could just bump the switcher for each milestone and only have major breaking milestones once a quarter or on some other cadence. | 1.0 | Add ability to reach older version documentation for PF4 - Documentation released that is out of date needs to be publicly accessible - this can be reached using a version switcher, archived somewhere (or in some other way). Users should be able to quickly identify which version the documentation refers to and view documentation for older releases.
Discussion is needed about whether we do this for major breaking changes (so when @patternfly/patternfly bumps from 2.x to 3.x - do we archive and allow people to switch back to 2.x) and then just update the 3.x visible documentation until we hit 4.x. We also need to discuss what this means for the @patternfly/react-* packages which can move independently. A version switcher for those packages could be harder to embed. We could just bump the switcher for each milestone and only have major breaking milestones once a quarter or on some other cadence. | infrastructure | add ability to reach older version documentation for documentation released that is out of date needs to be publicly accessible this can be reached using a version switcher archived somewhere or in some other way users should be able to quickly identify which version the documentation refers to and view documentation for older releases discussion is needed about whether we do this for major breaking changes so when patternfly patternfly bumps from x to x do we archive and allow people to switch back to x and then just update the x visible documentation until we hit x we also need to discuss what this means for the patternfly react packages which can move independently a version switcher for those packages could be harder to embed we could just bump the switcher for each milestone and only have major breaking milestones once a quarter or on some other cadence | 1 |
758,208 | 26,546,459,629 | IssuesEvent | 2023-01-20 00:56:05 | GoogleCloudPlatform/golang-samples | https://api.github.com/repos/GoogleCloudPlatform/golang-samples | closed | functions/security: Failure failed | type: bug priority: p1 api: cloudfunctions samples flakybot: issue | Note: #2013 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 271d0f45b7d9954f2a20aa00135bce551f86b5fc
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/c26f6e35-16a5-4069-a514-dc2323f1e02f), [Sponge](http://sponge2/c26f6e35-16a5-4069-a514-dc2323f1e02f)
status: failed
<details><summary>Test output</summary><br><pre>2023/01/18 18:59:32 Running: /tmp/google-cloud-sdk/bin/gcloud gcloud functions deploy MakeGetRequestCloudFunction-echo --entry-point=MakeGetRequestCloudFunction --runtime=go113 --no-allow-unauthenticated --project=golang-samples-tests-5 --trigger-http
2023/01/18 19:00:41 Deploying function (may take a while - up to 2 minutes)...
..
For Cloud Build Logs, visit: https://console.cloud.google.com/cloud-build/builds;region=us-central1/38c687aa-edc3-4a6d-8849-ca6cf0d9fc83?project=402919200380
..................................failed.
ERROR: (gcloud.functions.deploy) OperationError: code=3, message=Build failed: # golang.org/x/net/http2
/layers/google.go.functions-framework/gopath/pkg/mod/golang.org/x/net@v0.0.0-20221014081412-f15817d10f9b/http2/transport.go:437:45: undefined: os.ErrDeadlineExceeded
note: module requires Go 1.17; Error ID: 975560ac
2023/01/18 19:00:41 Setup: Deploy target function: exit status 1</pre></details> | 1.0 | functions/security: Failure failed - Note: #2013 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 271d0f45b7d9954f2a20aa00135bce551f86b5fc
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/c26f6e35-16a5-4069-a514-dc2323f1e02f), [Sponge](http://sponge2/c26f6e35-16a5-4069-a514-dc2323f1e02f)
status: failed
<details><summary>Test output</summary><br><pre>2023/01/18 18:59:32 Running: /tmp/google-cloud-sdk/bin/gcloud gcloud functions deploy MakeGetRequestCloudFunction-echo --entry-point=MakeGetRequestCloudFunction --runtime=go113 --no-allow-unauthenticated --project=golang-samples-tests-5 --trigger-http
2023/01/18 19:00:41 Deploying function (may take a while - up to 2 minutes)...
..
For Cloud Build Logs, visit: https://console.cloud.google.com/cloud-build/builds;region=us-central1/38c687aa-edc3-4a6d-8849-ca6cf0d9fc83?project=402919200380
..................................failed.
ERROR: (gcloud.functions.deploy) OperationError: code=3, message=Build failed: # golang.org/x/net/http2
/layers/google.go.functions-framework/gopath/pkg/mod/golang.org/x/net@v0.0.0-20221014081412-f15817d10f9b/http2/transport.go:437:45: undefined: os.ErrDeadlineExceeded
note: module requires Go 1.17; Error ID: 975560ac
2023/01/18 19:00:41 Setup: Deploy target function: exit status 1</pre></details> | non_infrastructure | functions security failure failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output running tmp google cloud sdk bin gcloud gcloud functions deploy makegetrequestcloudfunction echo entry point makegetrequestcloudfunction runtime no allow unauthenticated project golang samples tests trigger http deploying function may take a while up to minutes for cloud build logs visit failed error gcloud functions deploy operationerror code message build failed golang org x net layers google go functions framework gopath pkg mod golang org x net transport go undefined os errdeadlineexceeded note module requires go error id setup deploy target function exit status | 0 |
29,464 | 24,031,271,497 | IssuesEvent | 2022-09-15 15:12:11 | MissouriMRR/SUAS-2023 | https://api.github.com/repos/MissouriMRR/SUAS-2023 | opened | Flight-Vision Pipeline | enhancement infrastructure vision pipeline | # Issue Name
## Problem
Need ability for flight and vision to communicate.
## Solution
Need pipeline object that, given a flight state, will run vision code and return results to flight infrastructure.
| 1.0 | Flight-Vision Pipeline - # Issue Name
## Problem
Need ability for flight and vision to communicate.
## Solution
Need pipeline object that, given a flight state, will run vision code and return results to flight infrastructure.
| infrastructure | flight vision pipeline issue name problem need ability for flight and vision to communicate solution need pipeline object that given a flight state will run vision code and return results to flight infrastructure | 1 |
297,380 | 25,726,014,369 | IssuesEvent | 2022-12-07 16:42:44 | jvegax/decide-front | https://api.github.com/repos/jvegax/decide-front | opened | [TEST] - Detalles de una votacion | test 🧪 | Se debe comprobar que se muestra correctamente los detalles de una votación | 1.0 | [TEST] - Detalles de una votacion - Se debe comprobar que se muestra correctamente los detalles de una votación | non_infrastructure | detalles de una votacion se debe comprobar que se muestra correctamente los detalles de una votación | 0 |
32,803 | 27,004,815,285 | IssuesEvent | 2023-02-10 10:46:09 | Altinn/altinn-platform | https://api.github.com/repos/Altinn/altinn-platform | closed | Analyse use of Vertical Pod Autoscaling in Azure Kubernetes Service (AKS) | solution/apps area/infrastructure kind/analysis tm/no | ### Description
Azure have new VPA in preview.
Vertical Pod Autoscaler provides the following benefits:
- It analyzes and adjusts processor and memory resources to right size your applications. VPA isn't only responsible for scaling up, but also for scaling down based on their resource use over time.
- A Pod is evicted if it needs to change its resource requests if its scaling mode is set to auto or recreate.
- Set CPU and memory constraints for individual containers by specifying a resource policy
- Ensures nodes have correct resources for pod scheduling
- Configurable logging of any adjustments to processor or memory resources made
- Improve cluster resource utilization and frees up CPU and memory for other pods.
https://learn.microsoft.com/en-us/azure/aks/vertical-pod-autoscaler
### In scope
Turn on VPA in one AT environment and check how it handles the scaling.
### How to test
Turn on vpa with command:
```
az aks update -n myAKSCluster -g myResourceGroup --enable-vpa
```
Create vpa manifest for all running deployments:
```
#!/bin/bash
DEPLOYMENT_NAME=$(kubectl get deployment -oname | sed '/traefik/d' | sed 's/deployment.apps\///')
for i in $DEPLOYMENT_NAME;do
cat <<EOF | kubectl apply -f -
apiVersion: autoscaling.k8s.io/v1
kind: VerticalPodAutoscaler
metadata:
name: vpa-auto
spec:
targetRef:
apiVersion: "apps/v1"
kind: Deployment
name: "${i}"
updatePolicy:
updateMode: "Auto"
EOF
done
```
### Analysis
Check how cluster adjust resources for deployments.
### Conclusion
_No response_ | 1.0 | Analyse use of Vertical Pod Autoscaling in Azure Kubernetes Service (AKS) - ### Description
Azure have new VPA in preview.
Vertical Pod Autoscaler provides the following benefits:
- It analyzes and adjusts processor and memory resources to right size your applications. VPA isn't only responsible for scaling up, but also for scaling down based on their resource use over time.
- A Pod is evicted if it needs to change its resource requests if its scaling mode is set to auto or recreate.
- Set CPU and memory constraints for individual containers by specifying a resource policy
- Ensures nodes have correct resources for pod scheduling
- Configurable logging of any adjustments to processor or memory resources made
- Improve cluster resource utilization and frees up CPU and memory for other pods.
https://learn.microsoft.com/en-us/azure/aks/vertical-pod-autoscaler
### In scope
Turn on VPA in one AT environment and check how it handles the scaling.
### How to test
Turn on vpa with command:
```
az aks update -n myAKSCluster -g myResourceGroup --enable-vpa
```
Create vpa manifest for all running deployments:
```
#!/bin/bash
DEPLOYMENT_NAME=$(kubectl get deployment -oname | sed '/traefik/d' | sed 's/deployment.apps\///')
for i in $DEPLOYMENT_NAME;do
cat <<EOF | kubectl apply -f -
apiVersion: autoscaling.k8s.io/v1
kind: VerticalPodAutoscaler
metadata:
name: vpa-auto
spec:
targetRef:
apiVersion: "apps/v1"
kind: Deployment
name: "${i}"
updatePolicy:
updateMode: "Auto"
EOF
done
```
### Analysis
Check how cluster adjust resources for deployments.
### Conclusion
_No response_ | infrastructure | analyse use of vertical pod autoscaling in azure kubernetes service aks description azure have new vpa in preview vertical pod autoscaler provides the following benefits it analyzes and adjusts processor and memory resources to right size your applications vpa isn t only responsible for scaling up but also for scaling down based on their resource use over time a pod is evicted if it needs to change its resource requests if its scaling mode is set to auto or recreate set cpu and memory constraints for individual containers by specifying a resource policy ensures nodes have correct resources for pod scheduling configurable logging of any adjustments to processor or memory resources made improve cluster resource utilization and frees up cpu and memory for other pods in scope turn on vpa in one at environment and check how it handles the scaling how to test turn on vpa with command az aks update n myakscluster g myresourcegroup enable vpa create vpa manifest for all running deployments bin bash deployment name kubectl get deployment oname sed traefik d sed s deployment apps for i in deployment name do cat eof kubectl apply f apiversion autoscaling io kind verticalpodautoscaler metadata name vpa auto spec targetref apiversion apps kind deployment name i updatepolicy updatemode auto eof done analysis check how cluster adjust resources for deployments conclusion no response | 1 |
250,579 | 27,104,983,010 | IssuesEvent | 2023-02-15 11:24:49 | uriel-naor/Demo | https://api.github.com/repos/uriel-naor/Demo | closed | cryptography-2.0.2.tar.gz: 3 vulnerabilities (highest severity is: 9.1) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p></summary>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (cryptography version) | Fix PR available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-36242](https://www.mend.io/vulnerability-database/CVE-2020-36242) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | cryptography-2.0.2.tar.gz | Direct | N/A | ❌ |
| [CVE-2018-10903](https://www.mend.io/vulnerability-database/CVE-2018-10903) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | cryptography-2.0.2.tar.gz | Direct | 2.3 | ✅ |
| [CVE-2020-25659](https://www.mend.io/vulnerability-database/CVE-2020-25659) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | cryptography-2.0.2.tar.gz | Direct | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-36242</summary>
### Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-2.0.2.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In the cryptography package before 3.3.2 for Python, certain sequences of update calls to symmetrically encrypt multi-GB values could result in an integer overflow and buffer overflow, as demonstrated by the Fernet class.
<p>Publish Date: 2021-02-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36242>CVE-2020-36242</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-02-07</p>
<p>Fix Resolution: cryptography - 3.3.2</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-10903</summary>
### Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-2.0.2.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in python-cryptography versions between >=1.9.0 and <2.3. The finalize_with_tag API did not enforce a minimum tag length. If a user did not validate the input length prior to passing it to finalize_with_tag an attacker could craft an invalid payload with a shortened tag (e.g. 1 byte) such that they would have a 1 in 256 chance of passing the MAC check. GCM tag forgeries can cause key leakage.
<p>Publish Date: 2018-07-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-10903>CVE-2018-10903</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10903">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10903</a></p>
<p>Release Date: 2018-07-30</p>
<p>Fix Resolution (cryptography): 2.3</p>
<p>Direct dependency fix Resolution (cryptography): 2.3</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-25659</summary>
### Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-2.0.2.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
python-cryptography 3.2 is vulnerable to Bleichenbacher timing attacks in the RSA decryption API, via timed processing of valid PKCS#1 v1.5 ciphertext.
<p>Publish Date: 2021-01-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25659>CVE-2020-25659</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/pyca/cryptography/security/advisories/GHSA-hggm-jpg3-v476">https://github.com/pyca/cryptography/security/advisories/GHSA-hggm-jpg3-v476</a></p>
<p>Release Date: 2021-01-11</p>
<p>Fix Resolution: 3.2</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | cryptography-2.0.2.tar.gz: 3 vulnerabilities (highest severity is: 9.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p></summary>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (cryptography version) | Fix PR available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-36242](https://www.mend.io/vulnerability-database/CVE-2020-36242) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | cryptography-2.0.2.tar.gz | Direct | N/A | ❌ |
| [CVE-2018-10903](https://www.mend.io/vulnerability-database/CVE-2018-10903) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | cryptography-2.0.2.tar.gz | Direct | 2.3 | ✅ |
| [CVE-2020-25659](https://www.mend.io/vulnerability-database/CVE-2020-25659) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | cryptography-2.0.2.tar.gz | Direct | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-36242</summary>
### Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-2.0.2.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In the cryptography package before 3.3.2 for Python, certain sequences of update calls to symmetrically encrypt multi-GB values could result in an integer overflow and buffer overflow, as demonstrated by the Fernet class.
<p>Publish Date: 2021-02-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36242>CVE-2020-36242</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-02-07</p>
<p>Fix Resolution: cryptography - 3.3.2</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-10903</summary>
### Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-2.0.2.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in python-cryptography versions between >=1.9.0 and <2.3. The finalize_with_tag API did not enforce a minimum tag length. If a user did not validate the input length prior to passing it to finalize_with_tag an attacker could craft an invalid payload with a shortened tag (e.g. 1 byte) such that they would have a 1 in 256 chance of passing the MAC check. GCM tag forgeries can cause key leakage.
<p>Publish Date: 2018-07-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-10903>CVE-2018-10903</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10903">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10903</a></p>
<p>Release Date: 2018-07-30</p>
<p>Fix Resolution (cryptography): 2.3</p>
<p>Direct dependency fix Resolution (cryptography): 2.3</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-25659</summary>
### Vulnerable Library - <b>cryptography-2.0.2.tar.gz</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz">https://files.pythonhosted.org/packages/0a/23/ffee389a45f752dc745706cdb2d354b91ddb71594506fac45236cf67af49/cryptography-2.0.2.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-2.0.2.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uriel-naor/Demo/commit/421ca6d3a2f9c09f8e10d23469f88884889888a2">421ca6d3a2f9c09f8e10d23469f88884889888a2</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
python-cryptography 3.2 is vulnerable to Bleichenbacher timing attacks in the RSA decryption API, via timed processing of valid PKCS#1 v1.5 ciphertext.
<p>Publish Date: 2021-01-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25659>CVE-2020-25659</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/pyca/cryptography/security/advisories/GHSA-hggm-jpg3-v476">https://github.com/pyca/cryptography/security/advisories/GHSA-hggm-jpg3-v476</a></p>
<p>Release Date: 2021-01-11</p>
<p>Fix Resolution: 3.2</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_infrastructure | cryptography tar gz vulnerabilities highest severity is autoclosed vulnerable library cryptography tar gz cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt found in head commit a href vulnerabilities cve severity cvss dependency type fixed in cryptography version fix pr available high cryptography tar gz direct n a high cryptography tar gz direct medium cryptography tar gz direct n a details cve vulnerable library cryptography tar gz cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x cryptography tar gz vulnerable library found in head commit a href found in base branch main vulnerability details in the cryptography package before for python certain sequences of update calls to symmetrically encrypt multi gb values could result in an integer overflow and buffer overflow as demonstrated by the fernet class publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution cryptography in order to enable automatic remediation please create cve vulnerable library cryptography tar gz cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x cryptography tar gz vulnerable library found in head commit a href found in base branch main vulnerability details a flaw was found in python cryptography versions between and the finalize with tag api did not enforce a minimum tag length if a user did not validate the input length prior to passing it to finalize with tag an attacker could craft an invalid payload with a shortened tag e g byte such that they would have a in chance of passing the mac check gcm tag forgeries can cause key leakage publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cryptography direct dependency fix resolution cryptography rescue worker helmet automatic remediation is available for this issue cve vulnerable library cryptography tar gz cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x cryptography tar gz vulnerable library found in head commit a href found in base branch main vulnerability details python cryptography is vulnerable to bleichenbacher timing attacks in the rsa decryption api via timed processing of valid pkcs ciphertext publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution in order to enable automatic remediation please create rescue worker helmet automatic remediation is available for this issue | 0 |
551,765 | 16,188,729,542 | IssuesEvent | 2021-05-04 03:56:10 | remnoteio/remnote-issues | https://api.github.com/repos/remnoteio/remnote-issues | closed | Copy / pasting Rem doesn't preserve bullet formatting | checked editor-polish fixed-in-next-update fixed-in-remnote-1.3.7 priority=2 | **Describe the bug**
Copying rem seems to preserve most rem-level formatting, like header levels, but not bullet hiding.
**To Reproduce**
Steps to reproduce the behavior:
1. Create some Rem with bullets hidden
2. Copy the Rem (multiple Rem in my case)
3. Paste in another doc
**Expected behavior**
Expected bullets to be hidden when pasted.
**Screenshots**
If applicable, add screenshots to help explain your problem.
After pasting (bullets were hidden for all source rem)

**Desktop (please complete the following information):**
- OS: macOS
- Browser: Chrome or native app
- Version: Chrome 87 or Desktop app 1.1.10 | 1.0 | Copy / pasting Rem doesn't preserve bullet formatting - **Describe the bug**
Copying rem seems to preserve most rem-level formatting, like header levels, but not bullet hiding.
**To Reproduce**
Steps to reproduce the behavior:
1. Create some Rem with bullets hidden
2. Copy the Rem (multiple Rem in my case)
3. Paste in another doc
**Expected behavior**
Expected bullets to be hidden when pasted.
**Screenshots**
If applicable, add screenshots to help explain your problem.
After pasting (bullets were hidden for all source rem)

**Desktop (please complete the following information):**
- OS: macOS
- Browser: Chrome or native app
- Version: Chrome 87 or Desktop app 1.1.10 | non_infrastructure | copy pasting rem doesn t preserve bullet formatting describe the bug copying rem seems to preserve most rem level formatting like header levels but not bullet hiding to reproduce steps to reproduce the behavior create some rem with bullets hidden copy the rem multiple rem in my case paste in another doc expected behavior expected bullets to be hidden when pasted screenshots if applicable add screenshots to help explain your problem after pasting bullets were hidden for all source rem desktop please complete the following information os macos browser chrome or native app version chrome or desktop app | 0 |
345,410 | 30,809,057,210 | IssuesEvent | 2023-08-01 09:10:28 | status-im/status-go | https://api.github.com/repos/status-im/status-go | closed | TestTokenMasterAcceptMemberRequestToJoin failed | flaky test | relate error log:
```
2023-08-01T07:46:57.246Z DEBUG pushnotificationclient/client.go:713 handling message scheduled
communities_messenger_helpers_test.go:300:
Error Trace: /home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_messenger_helpers_test.go:300
/home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_messenger_helpers_test.go:314
/home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_events_utils_test.go:358
/home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_events_token_master_test.go:128
Error: Should be true
Test: TestTokenMasterCommunityEventsSuite/TestTokenMasterAcceptMemberRequestToJoin
```
cc @mprakhov | 1.0 | TestTokenMasterAcceptMemberRequestToJoin failed - relate error log:
```
2023-08-01T07:46:57.246Z DEBUG pushnotificationclient/client.go:713 handling message scheduled
communities_messenger_helpers_test.go:300:
Error Trace: /home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_messenger_helpers_test.go:300
/home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_messenger_helpers_test.go:314
/home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_events_utils_test.go:358
/home/jenkins/workspace/status-go_prs_tests_PR-3823/protocol/communities_events_token_master_test.go:128
Error: Should be true
Test: TestTokenMasterCommunityEventsSuite/TestTokenMasterAcceptMemberRequestToJoin
```
cc @mprakhov | non_infrastructure | testtokenmasteracceptmemberrequesttojoin failed relate error log debug pushnotificationclient client go handling message scheduled communities messenger helpers test go error trace home jenkins workspace status go prs tests pr protocol communities messenger helpers test go home jenkins workspace status go prs tests pr protocol communities messenger helpers test go home jenkins workspace status go prs tests pr protocol communities events utils test go home jenkins workspace status go prs tests pr protocol communities events token master test go error should be true test testtokenmastercommunityeventssuite testtokenmasteracceptmemberrequesttojoin cc mprakhov | 0 |
16,014 | 11,799,639,741 | IssuesEvent | 2020-03-18 16:12:36 | CodeForBaltimore/Bmore-Responsive | https://api.github.com/repos/CodeForBaltimore/Bmore-Responsive | opened | Convert to AWS Lambda and RDS | enhancement infrastructure | ### User Story
As a front-end developer
I'd like to access the API via AWS lambda
So that I can easily develop the application
### Acceptance Criteria
- [ ] Endpoints converted to Lambdas without sacrificing Express alternate functionality
- [ ] PR merged to `master`
### Notes
| 1.0 | Convert to AWS Lambda and RDS - ### User Story
As a front-end developer
I'd like to access the API via AWS lambda
So that I can easily develop the application
### Acceptance Criteria
- [ ] Endpoints converted to Lambdas without sacrificing Express alternate functionality
- [ ] PR merged to `master`
### Notes
| infrastructure | convert to aws lambda and rds user story as a front end developer i d like to access the api via aws lambda so that i can easily develop the application acceptance criteria endpoints converted to lambdas without sacrificing express alternate functionality pr merged to master notes | 1 |
316,104 | 27,137,041,851 | IssuesEvent | 2023-02-16 13:57:54 | zinclabs/zincobserve | https://api.github.com/repos/zinclabs/zincobserve | closed | update user | testing | ### Which Zinc Observe functionalities are relevant/related to the feature request?
users
### Description
User should be able to update the information
### Proposed solution
On user listing page, user will get the update icon where user will be able to change the detail.
### Alternatives considered
N/A | 1.0 | update user - ### Which Zinc Observe functionalities are relevant/related to the feature request?
users
### Description
User should be able to update the information
### Proposed solution
On user listing page, user will get the update icon where user will be able to change the detail.
### Alternatives considered
N/A | non_infrastructure | update user which zinc observe functionalities are relevant related to the feature request users description user should be able to update the information proposed solution on user listing page user will get the update icon where user will be able to change the detail alternatives considered n a | 0 |
188,767 | 15,170,356,919 | IssuesEvent | 2021-02-12 23:04:02 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | [suggestion][navigator 2.0] Make it a lot simpler for the end developer | customer: crowd d: api docs d: examples documentation f: routes framework passed first triage proposal | I just finished studying the Navigator 2.0 documentation and I have to say it is REALLY complex. I'm not saying it's badly done but it's very very complex, difficult to grasp and use. It's far from being simple and easy to implement so much so that I no longer have a feeling I'm working with Flutter. I fell in love with Flutter for its simplicity. I don't know how but I have a feeling that for the end developer, the implementation can be done a lot simpler. Only my 2 cents.
Disclaimer: I may be wrong and I'm rather representing junior-level Flutter devs. I have been developing in Flutter for 8 months I have created a few fully funcional apps (iOS/Android/web), some of them are published on my github. | 1.0 | [suggestion][navigator 2.0] Make it a lot simpler for the end developer - I just finished studying the Navigator 2.0 documentation and I have to say it is REALLY complex. I'm not saying it's badly done but it's very very complex, difficult to grasp and use. It's far from being simple and easy to implement so much so that I no longer have a feeling I'm working with Flutter. I fell in love with Flutter for its simplicity. I don't know how but I have a feeling that for the end developer, the implementation can be done a lot simpler. Only my 2 cents.
Disclaimer: I may be wrong and I'm rather representing junior-level Flutter devs. I have been developing in Flutter for 8 months I have created a few fully funcional apps (iOS/Android/web), some of them are published on my github. | non_infrastructure | make it a lot simpler for the end developer i just finished studying the navigator documentation and i have to say it is really complex i m not saying it s badly done but it s very very complex difficult to grasp and use it s far from being simple and easy to implement so much so that i no longer have a feeling i m working with flutter i fell in love with flutter for its simplicity i don t know how but i have a feeling that for the end developer the implementation can be done a lot simpler only my cents disclaimer i may be wrong and i m rather representing junior level flutter devs i have been developing in flutter for months i have created a few fully funcional apps ios android web some of them are published on my github | 0 |
27,138 | 21,193,914,548 | IssuesEvent | 2022-04-08 20:57:15 | firebase/firebase-ios-sdk | https://api.github.com/repos/firebase/firebase-ios-sdk | closed | FirebaseApp.app() logs an error when used as documented | api: core Infrastructure | FirebaseApp.app() is [documented](https://github.com/firebase/firebase-ios-sdk/blob/master/FirebaseCore/Sources/Public/FirebaseCore/FIRApp.h#L78) to "Returns the default app, or nil if the default app does not exist".
However, when used to check if the default app exists and it doesn't, it logs an [error](https://github.com/firebase/firebase-ios-sdk/blob/master/FirebaseCore/Sources/FIRApp.m#L251).
It's confusing to log an error when an API is used as documented. Perhaps, there should be separate APIs for getting the default app versus checking its existence? | 1.0 | FirebaseApp.app() logs an error when used as documented - FirebaseApp.app() is [documented](https://github.com/firebase/firebase-ios-sdk/blob/master/FirebaseCore/Sources/Public/FirebaseCore/FIRApp.h#L78) to "Returns the default app, or nil if the default app does not exist".
However, when used to check if the default app exists and it doesn't, it logs an [error](https://github.com/firebase/firebase-ios-sdk/blob/master/FirebaseCore/Sources/FIRApp.m#L251).
It's confusing to log an error when an API is used as documented. Perhaps, there should be separate APIs for getting the default app versus checking its existence? | infrastructure | firebaseapp app logs an error when used as documented firebaseapp app is to returns the default app or nil if the default app does not exist however when used to check if the default app exists and it doesn t it logs an it s confusing to log an error when an api is used as documented perhaps there should be separate apis for getting the default app versus checking its existence | 1 |
163,041 | 6,188,858,234 | IssuesEvent | 2017-07-04 11:20:40 | handsontable/handsontable | https://api.github.com/repos/handsontable/handsontable | opened | ESlint rules update | Core: handsontable Guess: few hours Priority: high Status: In development Type: Bug | ### Description
<!--- Tell us what happens and what should happen -->
To [prevent this bug](https://github.com/handsontable/handsontable/pull/4359) I suggest adding a rule to the ESLint. This rule will be responsible for checking if any of commits does not contain code which directly touches a `Handsontable` variable.
### Steps to reproduce
<!--- Provide steps to reproduce this issue -->
1. Build Handsontable example using browserify (or webpack etc.).
2. Open a `dropdown` editor.
3. There will be an error about trying to access to `Handsontable` variable which does not exist.
### Your environment
* Handsontable version: 0.32.0
| 1.0 | ESlint rules update - ### Description
<!--- Tell us what happens and what should happen -->
To [prevent this bug](https://github.com/handsontable/handsontable/pull/4359) I suggest adding a rule to the ESLint. This rule will be responsible for checking if any of commits does not contain code which directly touches a `Handsontable` variable.
### Steps to reproduce
<!--- Provide steps to reproduce this issue -->
1. Build Handsontable example using browserify (or webpack etc.).
2. Open a `dropdown` editor.
3. There will be an error about trying to access to `Handsontable` variable which does not exist.
### Your environment
* Handsontable version: 0.32.0
| non_infrastructure | eslint rules update description to i suggest adding a rule to the eslint this rule will be responsible for checking if any of commits does not contain code which directly touches a handsontable variable steps to reproduce build handsontable example using browserify or webpack etc open a dropdown editor there will be an error about trying to access to handsontable variable which does not exist your environment handsontable version | 0 |
33,390 | 4,478,697,452 | IssuesEvent | 2016-08-27 05:27:10 | brave/browser-laptop | https://api.github.com/repos/brave/browser-laptop | closed | update ledger welcome page with adjusted UI layout and partner logos | design ledger integration polish | 
showing some dims:

partner logos attached
[partner logos.zip](https://github.com/brave/browser-laptop/files/433649/partner.logos.zip)
| 1.0 | update ledger welcome page with adjusted UI layout and partner logos - 
showing some dims:

partner logos attached
[partner logos.zip](https://github.com/brave/browser-laptop/files/433649/partner.logos.zip)
| non_infrastructure | update ledger welcome page with adjusted ui layout and partner logos showing some dims partner logos attached | 0 |
3,571 | 4,414,460,548 | IssuesEvent | 2016-08-13 12:57:27 | MinetestForFun/server-minetestforfun | https://api.github.com/repos/MinetestForFun/server-minetestforfun | reopened | Update mtsatellite | Infrastructure Priority: Medium Website | Update mtsatellite to version 0.7 with the new tiles algo and less ressources needed :
- [x] https://bitbucket.org/s_l_teichmann/mtsatellite/commits/all
New automappercolor 10 times faster :
- [ ] https://bitbucket.org/s_l_teichmann/mtautocolors
| 1.0 | Update mtsatellite - Update mtsatellite to version 0.7 with the new tiles algo and less ressources needed :
- [x] https://bitbucket.org/s_l_teichmann/mtsatellite/commits/all
New automappercolor 10 times faster :
- [ ] https://bitbucket.org/s_l_teichmann/mtautocolors
| infrastructure | update mtsatellite update mtsatellite to version with the new tiles algo and less ressources needed new automappercolor times faster | 1 |
30,139 | 24,573,056,165 | IssuesEvent | 2022-10-13 10:07:02 | GitGuardian/ggshield | https://api.github.com/repos/GitGuardian/ggshield | opened | Release step fails to update Homebrew taps | bug status:: confirmed Infrastructure | ## Describe the bug
When releasing a new version of ggshield, the CI is supposed to update the Homebrew taps (https://github.com/GitGuardian/homebrew-tap and https://github.com/GitGuardian/homebrew-ggshield), but the update fails.
Steps to reproduce:
1. Release a new version of ggshield :/
Actual result:
The update does not find the new version. It prints a warning and falls back to the already published version:
```
poet -f ggshield \
| sed 's/Shiny new formula/Detect secrets in source code, scan your repos and docker images for leaks/g' \
| sed '7a\ license "MIT"' \
| tee homebrew-ggshield/Formula/ggshield.rb homebrew-tap/Formula/ggshield.rb
shell: /usr/bin/bash -e {0}
env:
pythonLocation: /opt/hostedtoolcache/Python/3.9.14/x64
LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.14/x64/lib
/opt/hostedtoolcache/Python/3.9.14/x64/lib/python3.9/site-packages/poet/poet.py:95: PackageVersionNotFoundWarning: Could not find an exact version match for ggshield version 1.13.5; using newest instead
```
(https://github.com/GitGuardian/ggshield/actions/runs/3236528564/jobs/5302492437)
Expected result:
No warning, and the update picks the right version.
## Preliminary investigation details
`poet` uses this URL `https://pypi.io/pypi/ggshield/json` to list available versions. The problem seems to be that at the time it accesses the URL, the new version is not listed yet.
A fix would be to query this URL in a loop until the new version is listed (with a limited number of retries and a delay between them), before calling `poet`.
| 1.0 | Release step fails to update Homebrew taps - ## Describe the bug
When releasing a new version of ggshield, the CI is supposed to update the Homebrew taps (https://github.com/GitGuardian/homebrew-tap and https://github.com/GitGuardian/homebrew-ggshield), but the update fails.
Steps to reproduce:
1. Release a new version of ggshield :/
Actual result:
The update does not find the new version. It prints a warning and falls back to the already published version:
```
poet -f ggshield \
| sed 's/Shiny new formula/Detect secrets in source code, scan your repos and docker images for leaks/g' \
| sed '7a\ license "MIT"' \
| tee homebrew-ggshield/Formula/ggshield.rb homebrew-tap/Formula/ggshield.rb
shell: /usr/bin/bash -e {0}
env:
pythonLocation: /opt/hostedtoolcache/Python/3.9.14/x64
LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.14/x64/lib
/opt/hostedtoolcache/Python/3.9.14/x64/lib/python3.9/site-packages/poet/poet.py:95: PackageVersionNotFoundWarning: Could not find an exact version match for ggshield version 1.13.5; using newest instead
```
(https://github.com/GitGuardian/ggshield/actions/runs/3236528564/jobs/5302492437)
Expected result:
No warning, and the update picks the right version.
## Preliminary investigation details
`poet` uses this URL `https://pypi.io/pypi/ggshield/json` to list available versions. The problem seems to be that at the time it accesses the URL, the new version is not listed yet.
A fix would be to query this URL in a loop until the new version is listed (with a limited number of retries and a delay between them), before calling `poet`.
| infrastructure | release step fails to update homebrew taps describe the bug when releasing a new version of ggshield the ci is supposed to update the homebrew taps and but the update fails steps to reproduce release a new version of ggshield actual result the update does not find the new version it prints a warning and falls back to the already published version poet f ggshield sed s shiny new formula detect secrets in source code scan your repos and docker images for leaks g sed license mit tee homebrew ggshield formula ggshield rb homebrew tap formula ggshield rb shell usr bin bash e env pythonlocation opt hostedtoolcache python ld library path opt hostedtoolcache python lib opt hostedtoolcache python lib site packages poet poet py packageversionnotfoundwarning could not find an exact version match for ggshield version using newest instead expected result no warning and the update picks the right version preliminary investigation details poet uses this url to list available versions the problem seems to be that at the time it accesses the url the new version is not listed yet a fix would be to query this url in a loop until the new version is listed with a limited number of retries and a delay between them before calling poet | 1 |
34,576 | 30,191,842,659 | IssuesEvent | 2023-07-04 16:03:31 | rusefi/rusefi | https://api.github.com/repos/rusefi/rusefi | closed | Let's admit it it's time to move to gradle | help wanted INFRASTRUCTURE maybe-offshore | Historically all java code was build by ant
At the moment we already have android app building using gradle so much of the code already builds under gradle and we have both and and gradle which is not optimal
we shall completely migrate from ant to gradle for all java code | 1.0 | Let's admit it it's time to move to gradle - Historically all java code was build by ant
At the moment we already have android app building using gradle so much of the code already builds under gradle and we have both and and gradle which is not optimal
we shall completely migrate from ant to gradle for all java code | infrastructure | let s admit it it s time to move to gradle historically all java code was build by ant at the moment we already have android app building using gradle so much of the code already builds under gradle and we have both and and gradle which is not optimal we shall completely migrate from ant to gradle for all java code | 1 |
599,486 | 18,274,938,955 | IssuesEvent | 2021-10-04 17:39:16 | gramaziokohler/integral_timber_joints | https://api.github.com/repos/gramaziokohler/integral_timber_joints | closed | Confirm TAMP for Nine Piece | Planning HighPriority | Make sure I can run planning and smoothing on Nine Piece and can visualize it back in Rhino.
This checks that
- Tools are defined okay.
- Actions and Movements and their ACM are defined okay.
- Planners and smoothers are working
- Resulting Trajectory can be read.
- Visualization is working. | 1.0 | Confirm TAMP for Nine Piece - Make sure I can run planning and smoothing on Nine Piece and can visualize it back in Rhino.
This checks that
- Tools are defined okay.
- Actions and Movements and their ACM are defined okay.
- Planners and smoothers are working
- Resulting Trajectory can be read.
- Visualization is working. | non_infrastructure | confirm tamp for nine piece make sure i can run planning and smoothing on nine piece and can visualize it back in rhino this checks that tools are defined okay actions and movements and their acm are defined okay planners and smoothers are working resulting trajectory can be read visualization is working | 0 |
205,231 | 7,094,889,358 | IssuesEvent | 2018-01-13 10:06:44 | dorukgezici/ITUscheduler | https://api.github.com/repos/dorukgezici/ITUscheduler | closed | Empty scheduleId after Save Schedule | bug top priority | Gives PAGE NOT FOUND error because redirected URL is http://ituscheduler.com/schedule/ (id missing). | 1.0 | Empty scheduleId after Save Schedule - Gives PAGE NOT FOUND error because redirected URL is http://ituscheduler.com/schedule/ (id missing). | non_infrastructure | empty scheduleid after save schedule gives page not found error because redirected url is id missing | 0 |
17,520 | 12,406,415,667 | IssuesEvent | 2020-05-21 19:05:52 | hicommonwealth/commonwealth-oss | https://api.github.com/repos/hicommonwealth/commonwealth-oss | opened | Remove/set timeout for controller refreshAll resets | infrastructure | ## Description
<!-- Ex. Is this a feature request, bug report, etc.? Describe the issue clearly. -->
When you enter and leave a community, the thread/comment/reaction/tag controllers are powered down and continually "reset" when you re-enter those communities. This leads to continuous calls to the API for data that we most likely had not too long ago.
We want to persist this data for at least some amount of time and simultaneously check its existence if we re-enter these communities. This way we actually cache content on the site, which it seems we don't really do for non-active chains or communities. | 1.0 | Remove/set timeout for controller refreshAll resets - ## Description
<!-- Ex. Is this a feature request, bug report, etc.? Describe the issue clearly. -->
When you enter and leave a community, the thread/comment/reaction/tag controllers are powered down and continually "reset" when you re-enter those communities. This leads to continuous calls to the API for data that we most likely had not too long ago.
We want to persist this data for at least some amount of time and simultaneously check its existence if we re-enter these communities. This way we actually cache content on the site, which it seems we don't really do for non-active chains or communities. | infrastructure | remove set timeout for controller refreshall resets description when you enter and leave a community the thread comment reaction tag controllers are powered down and continually reset when you re enter those communities this leads to continuous calls to the api for data that we most likely had not too long ago we want to persist this data for at least some amount of time and simultaneously check its existence if we re enter these communities this way we actually cache content on the site which it seems we don t really do for non active chains or communities | 1 |
30,904 | 25,153,814,652 | IssuesEvent | 2022-11-10 12:04:00 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | GC Reliability Framework CI job is broken on Ubuntu | area-GC-coreclr area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate | See https://ci.dot.net/job/dotnet_coreclr/job/master/job/x64_release_ubuntu_gc_reliability_framework_tst_prtest/1/console:
```
09:14:57 Unhandled Exception: System.ArgumentOutOfRangeException: Length cannot be less than zero.
09:14:57 Parameter name: length
09:14:57 at System.String.Substring(Int32 startIndex, Int32 length)
09:14:57 at ReliabilityFramework.Main(String[] args)
```
Doesn't repro locally for me, so if I were to hazard a guess it's that the CI is missing an environment variable that the RF is expecting to be present.
| 1.0 | GC Reliability Framework CI job is broken on Ubuntu - See https://ci.dot.net/job/dotnet_coreclr/job/master/job/x64_release_ubuntu_gc_reliability_framework_tst_prtest/1/console:
```
09:14:57 Unhandled Exception: System.ArgumentOutOfRangeException: Length cannot be less than zero.
09:14:57 Parameter name: length
09:14:57 at System.String.Substring(Int32 startIndex, Int32 length)
09:14:57 at ReliabilityFramework.Main(String[] args)
```
Doesn't repro locally for me, so if I were to hazard a guess it's that the CI is missing an environment variable that the RF is expecting to be present.
| infrastructure | gc reliability framework ci job is broken on ubuntu see unhandled exception system argumentoutofrangeexception length cannot be less than zero parameter name length at system string substring startindex length at reliabilityframework main string args doesn t repro locally for me so if i were to hazard a guess it s that the ci is missing an environment variable that the rf is expecting to be present | 1 |
32,842 | 27,032,007,440 | IssuesEvent | 2023-02-12 10:14:34 | FTBTeam/FTB-Mods-Issues | https://api.github.com/repos/FTBTeam/FTB-Mods-Issues | closed | [Feature Request]: Add ftblibrary to the dependencies on Curseforge | FTB Industrial Contraptions 3: Infrastructure | ### Mod
FTB Industrical Contraptions
### Description
When looking at the dependencies for FTB IC on Curseforge it only shows Architecture and Myrtrees, even though it also required FTB Library to run.
While FTB IC is set as a dependent of FTB Library on its Curseforge page this isn't as useful as I would have know to search for FTB Library ahead of time to check if it is a require dependancy.
### Describe the solution you'd like
Add FTB Library to the list of dependancies of FTB Industrial Contraptions.
### Describe alternatives you've considered
_No response_ | 1.0 | [Feature Request]: Add ftblibrary to the dependencies on Curseforge - ### Mod
FTB Industrical Contraptions
### Description
When looking at the dependencies for FTB IC on Curseforge it only shows Architecture and Myrtrees, even though it also required FTB Library to run.
While FTB IC is set as a dependent of FTB Library on its Curseforge page this isn't as useful as I would have know to search for FTB Library ahead of time to check if it is a require dependancy.
### Describe the solution you'd like
Add FTB Library to the list of dependancies of FTB Industrial Contraptions.
### Describe alternatives you've considered
_No response_ | infrastructure | add ftblibrary to the dependencies on curseforge mod ftb industrical contraptions description when looking at the dependencies for ftb ic on curseforge it only shows architecture and myrtrees even though it also required ftb library to run while ftb ic is set as a dependent of ftb library on its curseforge page this isn t as useful as i would have know to search for ftb library ahead of time to check if it is a require dependancy describe the solution you d like add ftb library to the list of dependancies of ftb industrial contraptions describe alternatives you ve considered no response | 1 |
184,454 | 31,926,107,985 | IssuesEvent | 2023-09-19 01:58:18 | AppFlowy-IO/AppFlowy | https://api.github.com/repos/AppFlowy-IO/AppFlowy | closed | [FR] End time for date field in grid/kanban | new feature needs design calendar | ### Description
Currently, the date/time that can be inputted in a date field only allows one time point. It would be better if a user can specify an end time as well. Requires adding the field in the backend and alterations in the UI.
### Impact
Users who would like to input date ranges for an event will be able to do so.
### Additional Context
_No response_ | 1.0 | [FR] End time for date field in grid/kanban - ### Description
Currently, the date/time that can be inputted in a date field only allows one time point. It would be better if a user can specify an end time as well. Requires adding the field in the backend and alterations in the UI.
### Impact
Users who would like to input date ranges for an event will be able to do so.
### Additional Context
_No response_ | non_infrastructure | end time for date field in grid kanban description currently the date time that can be inputted in a date field only allows one time point it would be better if a user can specify an end time as well requires adding the field in the backend and alterations in the ui impact users who would like to input date ranges for an event will be able to do so additional context no response | 0 |
172,848 | 21,055,576,317 | IssuesEvent | 2022-04-01 02:38:31 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [Security Solution]comma separated process.arg not wraps properly | bug triage_needed Team: SecuritySolution | **Describe the bug**
comma separated process.arg not wraps properly
**Build Details**
```
Version:8.2.0-SNAPSHOT
BUILD 51431
COMMIT a743498436a863e142592cb535b43f44c448851a
```
**Steps**
- Login to Kibana
- Generate some alert data , in our case we create a custom rule for process.name: "cmd.exe" and executed mutiple instance of cmd on windows host
- Click on Alert Flyout
- Observed that comma separated process.arg not wraps properly
**Screen-Shoot**



**Additional Details:**
- **_actual content copied in clipboard: process.args_**: "cmd,/c,rmdir,C:\Users\zeus\AppData\Local\Temp\peazip-tmp\.pztmp\neutral22033117,/s,/q"
- **_filter in of above process.args_**

| True | [Security Solution]comma separated process.arg not wraps properly - **Describe the bug**
comma separated process.arg not wraps properly
**Build Details**
```
Version:8.2.0-SNAPSHOT
BUILD 51431
COMMIT a743498436a863e142592cb535b43f44c448851a
```
**Steps**
- Login to Kibana
- Generate some alert data , in our case we create a custom rule for process.name: "cmd.exe" and executed mutiple instance of cmd on windows host
- Click on Alert Flyout
- Observed that comma separated process.arg not wraps properly
**Screen-Shoot**



**Additional Details:**
- **_actual content copied in clipboard: process.args_**: "cmd,/c,rmdir,C:\Users\zeus\AppData\Local\Temp\peazip-tmp\.pztmp\neutral22033117,/s,/q"
- **_filter in of above process.args_**

| non_infrastructure | comma separated process arg not wraps properly describe the bug comma separated process arg not wraps properly build details version snapshot build commit steps login to kibana generate some alert data in our case we create a custom rule for process name cmd exe and executed mutiple instance of cmd on windows host click on alert flyout observed that comma separated process arg not wraps properly screen shoot additional details actual content copied in clipboard process args cmd c rmdir c users zeus appdata local temp peazip tmp pztmp s q filter in of above process args | 0 |
26,204 | 4,614,643,755 | IssuesEvent | 2016-09-25 17:56:57 | ncarthy/kodi-crestron | https://api.github.com/repos/ncarthy/kodi-crestron | closed | Will no longer connect to XBMC | auto-migrated Priority-Medium Type-Defect | ```
The only change I can think of that I made recently was installing the Crestron
Studio software. At some point after doing that, I wanted to make a change to
my system. After compiling and uploading I noticed I could no longer control
my XBMC. So I loaded the pre-compiled version from the download (xbmc-test)
and it connected fine.
I was using version .6 but updated to the latest 1.01beta to see if that would
fix it.
The error I'm getting in debugger anytime I try to send a command is "Not
connected to XBMC at 192.168.1.201. Socket Error number is -1."
The processor is a Prodigy PMC3, my machine is Win7 x64, and I have the latest
version of all Crestron software.
Have you seen this error before?
```
Original issue reported on code.google.com by `matthews...@gmail.com` on 29 Dec 2012 at 2:04 | 1.0 | Will no longer connect to XBMC - ```
The only change I can think of that I made recently was installing the Crestron
Studio software. At some point after doing that, I wanted to make a change to
my system. After compiling and uploading I noticed I could no longer control
my XBMC. So I loaded the pre-compiled version from the download (xbmc-test)
and it connected fine.
I was using version .6 but updated to the latest 1.01beta to see if that would
fix it.
The error I'm getting in debugger anytime I try to send a command is "Not
connected to XBMC at 192.168.1.201. Socket Error number is -1."
The processor is a Prodigy PMC3, my machine is Win7 x64, and I have the latest
version of all Crestron software.
Have you seen this error before?
```
Original issue reported on code.google.com by `matthews...@gmail.com` on 29 Dec 2012 at 2:04 | non_infrastructure | will no longer connect to xbmc the only change i can think of that i made recently was installing the crestron studio software at some point after doing that i wanted to make a change to my system after compiling and uploading i noticed i could no longer control my xbmc so i loaded the pre compiled version from the download xbmc test and it connected fine i was using version but updated to the latest to see if that would fix it the error i m getting in debugger anytime i try to send a command is not connected to xbmc at socket error number is the processor is a prodigy my machine is and i have the latest version of all crestron software have you seen this error before original issue reported on code google com by matthews gmail com on dec at | 0 |
11,313 | 9,099,305,245 | IssuesEvent | 2019-02-20 03:48:31 | PuzzleServer/mainpuzzleserver | https://api.github.com/repos/PuzzleServer/mainpuzzleserver | closed | Add [Required] attribute to DB properties | infrastructure | If a property on an Entity Framework table is a nullable type (such as ```String``` or another class), the column is treated as optional and the DB will accept null. For values that shouldn't be null, we should set the ```[Required]``` attribute. | 1.0 | Add [Required] attribute to DB properties - If a property on an Entity Framework table is a nullable type (such as ```String``` or another class), the column is treated as optional and the DB will accept null. For values that shouldn't be null, we should set the ```[Required]``` attribute. | infrastructure | add attribute to db properties if a property on an entity framework table is a nullable type such as string or another class the column is treated as optional and the db will accept null for values that shouldn t be null we should set the attribute | 1 |
25,249 | 18,343,824,843 | IssuesEvent | 2021-10-08 01:40:57 | Samsung/thorvg | https://api.github.com/repos/Samsung/thorvg | closed | Add build test for windows | infrastructure | It would be better if we verify the windows build test using git action. | 1.0 | Add build test for windows - It would be better if we verify the windows build test using git action. | infrastructure | add build test for windows it would be better if we verify the windows build test using git action | 1 |
9,714 | 8,123,911,829 | IssuesEvent | 2018-08-16 15:51:44 | DMPRoadmap/roadmap | https://api.github.com/repos/DMPRoadmap/roadmap | closed | Create documentation for tests | effort-large infrastructure | For external contributors guide, create documentation for tests based on current strategy. Devs can discuss future/ideal strategy but in the interim we need something that enables folks like DMP OPIDoR to contribute effectively. | 1.0 | Create documentation for tests - For external contributors guide, create documentation for tests based on current strategy. Devs can discuss future/ideal strategy but in the interim we need something that enables folks like DMP OPIDoR to contribute effectively. | infrastructure | create documentation for tests for external contributors guide create documentation for tests based on current strategy devs can discuss future ideal strategy but in the interim we need something that enables folks like dmp opidor to contribute effectively | 1 |
16,883 | 12,152,147,199 | IssuesEvent | 2020-04-24 21:31:01 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Re-design playbook for grabbing evacuated pods due to NodeNotReady or similar | Infrastructure closed | https://trello.com/c/ilezv9cc/34-re-design-playbook-for-grabbing-evacuated-pods-due-to-nodenotready-or-similar
Legacy playbook written for ESIT on-call and other stakeholders to grab a list of affected pods when a NodeNotReady or similar event forces all pods on a node to be evicted elsewhere. Results produced suitable for cut/pasting into Slack #alerts, as well as feeding a suitable internal Incident ticket.
This broke when OpenShift upgraded to 3.10 since the data source went from the OS journal data to a containerized pod. Different data source, different options and ability to narrow the log results returned. | 1.0 | Re-design playbook for grabbing evacuated pods due to NodeNotReady or similar - https://trello.com/c/ilezv9cc/34-re-design-playbook-for-grabbing-evacuated-pods-due-to-nodenotready-or-similar
Legacy playbook written for ESIT on-call and other stakeholders to grab a list of affected pods when a NodeNotReady or similar event forces all pods on a node to be evicted elsewhere. Results produced suitable for cut/pasting into Slack #alerts, as well as feeding a suitable internal Incident ticket.
This broke when OpenShift upgraded to 3.10 since the data source went from the OS journal data to a containerized pod. Different data source, different options and ability to narrow the log results returned. | infrastructure | re design playbook for grabbing evacuated pods due to nodenotready or similar legacy playbook written for esit on call and other stakeholders to grab a list of affected pods when a nodenotready or similar event forces all pods on a node to be evicted elsewhere results produced suitable for cut pasting into slack alerts as well as feeding a suitable internal incident ticket this broke when openshift upgraded to since the data source went from the os journal data to a containerized pod different data source different options and ability to narrow the log results returned | 1 |
17,754 | 12,534,929,573 | IssuesEvent | 2020-06-04 20:22:40 | badoo/Reaktive | https://api.github.com/repos/badoo/Reaktive | closed | Flaky CI "Set up cache" step | Infrastructure | The step "Set up cache" is flaky. If it fails then rerunning it will fail all jobs.
Latest example: https://github.com/badoo/Reaktive/runs/539845234 | 1.0 | Flaky CI "Set up cache" step - The step "Set up cache" is flaky. If it fails then rerunning it will fail all jobs.
Latest example: https://github.com/badoo/Reaktive/runs/539845234 | infrastructure | flaky ci set up cache step the step set up cache is flaky if it fails then rerunning it will fail all jobs latest example | 1 |
40,677 | 5,312,693,385 | IssuesEvent | 2017-02-13 09:52:18 | davidbradburn/CFApp | https://api.github.com/repos/davidbradburn/CFApp | closed | Apprentice Graduation | CF Test | I would like to be able to graduate a apprentice resource to Level 12 but to have captured that the resource was an apprentice | 1.0 | Apprentice Graduation - I would like to be able to graduate a apprentice resource to Level 12 but to have captured that the resource was an apprentice | non_infrastructure | apprentice graduation i would like to be able to graduate a apprentice resource to level but to have captured that the resource was an apprentice | 0 |
32,862 | 27,045,209,529 | IssuesEvent | 2023-02-13 09:16:46 | statice/anonymeter | https://api.github.com/repos/statice/anonymeter | opened | Update CI with style checks | infrastructure | Integrate the CI with style checks in case more people want to contribute code. | 1.0 | Update CI with style checks - Integrate the CI with style checks in case more people want to contribute code. | infrastructure | update ci with style checks integrate the ci with style checks in case more people want to contribute code | 1 |
13,401 | 5,356,074,430 | IssuesEvent | 2017-02-20 14:45:59 | docker/docker | https://api.github.com/repos/docker/docker | opened | --cache-from does not cache last step | area/builder version/1.13 | <!--
If you are reporting a new issue, make sure that we do not have any duplicates
already open. You can ensure this by searching the issue list for this
repository. If there is a duplicate, please close your issue and add a comment
to the existing issue instead.
If you suspect your issue is a bug, please edit your issue description to
include the BUG REPORT INFORMATION shown below. If you fail to provide this
information within 7 days, we cannot debug your issue and will close it. We
will, however, reopen it if you later provide the information.
For more information about reporting issues, see
https://github.com/docker/docker/blob/master/CONTRIBUTING.md#reporting-other-issues
---------------------------------------------------
GENERAL SUPPORT INFORMATION
---------------------------------------------------
The GitHub issue tracker is for bug reports and feature requests.
General support can be found at the following locations:
- Docker Support Forums - https://forums.docker.com
- IRC - irc.freenode.net #docker channel
- Post a question on StackOverflow, using the Docker tag
---------------------------------------------------
BUG REPORT INFORMATION
---------------------------------------------------
Use the commands below to provide key information from your environment:
You do NOT have to include this information if this is a FEATURE REQUEST
-->
**Description**
Using `--cache-from` seems to ignore the last cached image layer in certain cases. see repro below.
**Steps to reproduce the issue:**
```sh
# Dockerfile
$ cat Dockerfile
FROM alpine:latest
RUN echo "test"
ENTRYPOINT ["sh"]
# first run, nothing cached
$ docker build -f Dockerfile . -t something
Sending build context to Docker daemon 14.85 kB
Step 1/3 : FROM alpine:latest
---> 88e169ea8f46
Step 2/3 : RUN echo "test"
---> Running in 88967ec466ca
test
---> 1f1ea7b19579
Removing intermediate container 88967ec466ca
Step 3/3 : ENTRYPOINT sh
---> Running in d9498875cf40
---> e3a30a317f03
Removing intermediate container d9498875cf40
Successfully built e3a30a317f03
# second run, everything cached
$ docker build -f Dockerfile .
Sending build context to Docker daemon 14.85 kB
Step 1/3 : FROM alpine:latest
---> 88e169ea8f46
Step 2/3 : RUN echo "test"
---> Using cache
---> 1f1ea7b19579
Step 3/3 : ENTRYPOINT sh
---> Using cache
---> e3a30a317f03
Successfully built e3a30a317f03
# third run, use --cache from - last step is not cached
$ docker build --cache-from something -f Dockerfile .
Sending build context to Docker daemon 14.85 kB
Step 1/3 : FROM alpine:latest
---> 88e169ea8f46
Step 2/3 : RUN echo "test"
---> Using cache
---> 1f1ea7b19579
Step 3/3 : ENTRYPOINT sh
---> Running in 92dd22c70bde
---> b0de7181ed64
Removing intermediate container 92dd22c70bde
Successfully built b0de7181ed64
```
**Describe the results you received:**
Using `--cache-from` will use all cached layers from the target image
**Describe the results you expected:**
It does not!
**Additional information you deem important (e.g. issue happens only occasionally):**
**Output of `docker version`:**
```
Client:
Version: 1.13.1
API version: 1.26
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: darwin/amd64
Server:
Version: 1.13.1
API version: 1.26 (minimum version 1.12)
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: linux/amd64
Experimental: true
```
**Output of `docker info`:**
```
Containers: 4
Running: 0
Paused: 0
Stopped: 4
Images: 5
Server Version: 1.13.1
Storage Driver: aufs
Root Dir: /var/lib/docker/aufs
Backing Filesystem: extfs
Dirs: 18
Dirperm1 Supported: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host ipvlan macvlan null overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1
runc version: 9df8b306d01f59d3a8029be411de015b7304dd8f
init version: 949e6fa
Security Options:
seccomp
Profile: default
Kernel Version: 4.9.8-moby
Operating System: Alpine Linux v3.5
OSType: linux
Architecture: x86_64
CPUs: 4
Total Memory: 1.952 GiB
Name: moby
ID: KKRA:W5WK:ES3L:FUCS:EEEP:LAND:I67M:E2TP:XBCI:OLN2:CC6Z:BR2Z
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): true
File Descriptors: 19
Goroutines: 27
System Time: 2017-02-20T14:39:25.139758208Z
EventsListeners: 1
No Proxy: *.local, 169.254/16
Registry: https://index.docker.io/v1/
Experimental: true
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
| 1.0 | --cache-from does not cache last step - <!--
If you are reporting a new issue, make sure that we do not have any duplicates
already open. You can ensure this by searching the issue list for this
repository. If there is a duplicate, please close your issue and add a comment
to the existing issue instead.
If you suspect your issue is a bug, please edit your issue description to
include the BUG REPORT INFORMATION shown below. If you fail to provide this
information within 7 days, we cannot debug your issue and will close it. We
will, however, reopen it if you later provide the information.
For more information about reporting issues, see
https://github.com/docker/docker/blob/master/CONTRIBUTING.md#reporting-other-issues
---------------------------------------------------
GENERAL SUPPORT INFORMATION
---------------------------------------------------
The GitHub issue tracker is for bug reports and feature requests.
General support can be found at the following locations:
- Docker Support Forums - https://forums.docker.com
- IRC - irc.freenode.net #docker channel
- Post a question on StackOverflow, using the Docker tag
---------------------------------------------------
BUG REPORT INFORMATION
---------------------------------------------------
Use the commands below to provide key information from your environment:
You do NOT have to include this information if this is a FEATURE REQUEST
-->
**Description**
Using `--cache-from` seems to ignore the last cached image layer in certain cases. see repro below.
**Steps to reproduce the issue:**
```sh
# Dockerfile
$ cat Dockerfile
FROM alpine:latest
RUN echo "test"
ENTRYPOINT ["sh"]
# first run, nothing cached
$ docker build -f Dockerfile . -t something
Sending build context to Docker daemon 14.85 kB
Step 1/3 : FROM alpine:latest
---> 88e169ea8f46
Step 2/3 : RUN echo "test"
---> Running in 88967ec466ca
test
---> 1f1ea7b19579
Removing intermediate container 88967ec466ca
Step 3/3 : ENTRYPOINT sh
---> Running in d9498875cf40
---> e3a30a317f03
Removing intermediate container d9498875cf40
Successfully built e3a30a317f03
# second run, everything cached
$ docker build -f Dockerfile .
Sending build context to Docker daemon 14.85 kB
Step 1/3 : FROM alpine:latest
---> 88e169ea8f46
Step 2/3 : RUN echo "test"
---> Using cache
---> 1f1ea7b19579
Step 3/3 : ENTRYPOINT sh
---> Using cache
---> e3a30a317f03
Successfully built e3a30a317f03
# third run, use --cache from - last step is not cached
$ docker build --cache-from something -f Dockerfile .
Sending build context to Docker daemon 14.85 kB
Step 1/3 : FROM alpine:latest
---> 88e169ea8f46
Step 2/3 : RUN echo "test"
---> Using cache
---> 1f1ea7b19579
Step 3/3 : ENTRYPOINT sh
---> Running in 92dd22c70bde
---> b0de7181ed64
Removing intermediate container 92dd22c70bde
Successfully built b0de7181ed64
```
**Describe the results you received:**
Using `--cache-from` will use all cached layers from the target image
**Describe the results you expected:**
It does not!
**Additional information you deem important (e.g. issue happens only occasionally):**
**Output of `docker version`:**
```
Client:
Version: 1.13.1
API version: 1.26
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: darwin/amd64
Server:
Version: 1.13.1
API version: 1.26 (minimum version 1.12)
Go version: go1.7.5
Git commit: 092cba3
Built: Wed Feb 8 08:47:51 2017
OS/Arch: linux/amd64
Experimental: true
```
**Output of `docker info`:**
```
Containers: 4
Running: 0
Paused: 0
Stopped: 4
Images: 5
Server Version: 1.13.1
Storage Driver: aufs
Root Dir: /var/lib/docker/aufs
Backing Filesystem: extfs
Dirs: 18
Dirperm1 Supported: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host ipvlan macvlan null overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1
runc version: 9df8b306d01f59d3a8029be411de015b7304dd8f
init version: 949e6fa
Security Options:
seccomp
Profile: default
Kernel Version: 4.9.8-moby
Operating System: Alpine Linux v3.5
OSType: linux
Architecture: x86_64
CPUs: 4
Total Memory: 1.952 GiB
Name: moby
ID: KKRA:W5WK:ES3L:FUCS:EEEP:LAND:I67M:E2TP:XBCI:OLN2:CC6Z:BR2Z
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): true
File Descriptors: 19
Goroutines: 27
System Time: 2017-02-20T14:39:25.139758208Z
EventsListeners: 1
No Proxy: *.local, 169.254/16
Registry: https://index.docker.io/v1/
Experimental: true
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
| non_infrastructure | cache from does not cache last step if you are reporting a new issue make sure that we do not have any duplicates already open you can ensure this by searching the issue list for this repository if there is a duplicate please close your issue and add a comment to the existing issue instead if you suspect your issue is a bug please edit your issue description to include the bug report information shown below if you fail to provide this information within days we cannot debug your issue and will close it we will however reopen it if you later provide the information for more information about reporting issues see general support information the github issue tracker is for bug reports and feature requests general support can be found at the following locations docker support forums irc irc freenode net docker channel post a question on stackoverflow using the docker tag bug report information use the commands below to provide key information from your environment you do not have to include this information if this is a feature request description using cache from seems to ignore the last cached image layer in certain cases see repro below steps to reproduce the issue sh dockerfile cat dockerfile from alpine latest run echo test entrypoint first run nothing cached docker build f dockerfile t something sending build context to docker daemon kb step from alpine latest step run echo test running in test removing intermediate container step entrypoint sh running in removing intermediate container successfully built second run everything cached docker build f dockerfile sending build context to docker daemon kb step from alpine latest step run echo test using cache step entrypoint sh using cache successfully built third run use cache from last step is not cached docker build cache from something f dockerfile sending build context to docker daemon kb step from alpine latest step run echo test using cache step entrypoint sh running in removing intermediate container successfully built describe the results you received using cache from will use all cached layers from the target image describe the results you expected it does not additional information you deem important e g issue happens only occasionally output of docker version client version api version go version git commit built wed feb os arch darwin server version api version minimum version go version git commit built wed feb os arch linux experimental true output of docker info containers running paused stopped images server version storage driver aufs root dir var lib docker aufs backing filesystem extfs dirs supported true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host ipvlan macvlan null overlay swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options seccomp profile default kernel version moby operating system alpine linux ostype linux architecture cpus total memory gib name moby id kkra fucs eeep land xbci docker root dir var lib docker debug mode client false debug mode server true file descriptors goroutines system time eventslisteners no proxy local registry experimental true insecure registries live restore enabled false additional environment details aws virtualbox physical etc | 0 |
14,715 | 11,088,988,701 | IssuesEvent | 2019-12-14 15:19:12 | ForNeVeR/wpf-math | https://api.github.com/repos/ForNeVeR/wpf-math | closed | Add TFM into nuspec | kind:infrastructure | See [here](https://docs.microsoft.com/en-us/nuget/reference/errors-and-warnings/nu5128): we could probably either use `dotnet pack` for nuget package generation or update our manually-written nuspec to include that.
Also, it would be nice to generate the package on the CI server, and to automatically upload it to GitHub. | 1.0 | Add TFM into nuspec - See [here](https://docs.microsoft.com/en-us/nuget/reference/errors-and-warnings/nu5128): we could probably either use `dotnet pack` for nuget package generation or update our manually-written nuspec to include that.
Also, it would be nice to generate the package on the CI server, and to automatically upload it to GitHub. | infrastructure | add tfm into nuspec see we could probably either use dotnet pack for nuget package generation or update our manually written nuspec to include that also it would be nice to generate the package on the ci server and to automatically upload it to github | 1 |
33,117 | 27,241,564,187 | IssuesEvent | 2023-02-21 20:57:43 | apache/buildstream | https://api.github.com/repos/apache/buildstream | closed | Publish to pypi from CI | infrastructure | [See original issue on GitLab](https://gitlab.com/BuildStream/buildstream/-/issues/1282)
In GitLab by [[Gitlab user @jjardon]](https://gitlab.com/jjardon) on Apr 2, 2020, 10:52
AFAIK publishing new releases at https://pypi.org/project/BuildStream/ is a manual process | 1.0 | Publish to pypi from CI - [See original issue on GitLab](https://gitlab.com/BuildStream/buildstream/-/issues/1282)
In GitLab by [[Gitlab user @jjardon]](https://gitlab.com/jjardon) on Apr 2, 2020, 10:52
AFAIK publishing new releases at https://pypi.org/project/BuildStream/ is a manual process | infrastructure | publish to pypi from ci in gitlab by on apr afaik publishing new releases at is a manual process | 1 |
19,002 | 13,185,164,213 | IssuesEvent | 2020-08-12 20:50:58 | ustaxcourt/ef-cms | https://api.github.com/repos/ustaxcourt/ef-cms | closed | Configure technical monitoring and alerting. | infrastructure | As the Court’s operating team, so that I can ensure EF-CMS is operating correctly, I need to be able to monitor the performance of the application’s components and be alerted to issues.
Acceptance criteria
- Notifications are sent when:
- The application is partially or completely down
- There are usage patterns which indicate a security breach
- There is a spike in traffic which may cause unusually high AWS billing
- Routine processes are not occurring
- Elasticsearch is overloaded | 1.0 | Configure technical monitoring and alerting. - As the Court’s operating team, so that I can ensure EF-CMS is operating correctly, I need to be able to monitor the performance of the application’s components and be alerted to issues.
Acceptance criteria
- Notifications are sent when:
- The application is partially or completely down
- There are usage patterns which indicate a security breach
- There is a spike in traffic which may cause unusually high AWS billing
- Routine processes are not occurring
- Elasticsearch is overloaded | infrastructure | configure technical monitoring and alerting as the court’s operating team so that i can ensure ef cms is operating correctly i need to be able to monitor the performance of the application’s components and be alerted to issues acceptance criteria notifications are sent when the application is partially or completely down there are usage patterns which indicate a security breach there is a spike in traffic which may cause unusually high aws billing routine processes are not occurring elasticsearch is overloaded | 1 |
14,191 | 10,697,213,828 | IssuesEvent | 2019-10-23 16:04:03 | OregonDigital/OD2 | https://api.github.com/repos/OregonDigital/OD2 | closed | WC7 - Setup on-premises k8s cluster | Epic Infrastructure Priority - High | During WC7 we need to build out the new on-premises staging environment at OSU.
- [x] Setup control plane and worker VMs
- [x] Setup Kubernetes cluster
- [x] Integrate with OSULP Rancher
- [x] Update CI/CD build/deploy pipeline
- [x] Provision storage
- [x] Test OD2 in new staging environment
**lib-odcp1-3** (control plane nodes)
- 4 CPU
- 8GB RAM
- 100GB disk
**lib-odwork1-4** (worker nodes)
- 4 CPU
- 24GB RAM
- 100GB disk
**32 TB NFS Volume** (high capacity asset/derivative storage) | 1.0 | WC7 - Setup on-premises k8s cluster - During WC7 we need to build out the new on-premises staging environment at OSU.
- [x] Setup control plane and worker VMs
- [x] Setup Kubernetes cluster
- [x] Integrate with OSULP Rancher
- [x] Update CI/CD build/deploy pipeline
- [x] Provision storage
- [x] Test OD2 in new staging environment
**lib-odcp1-3** (control plane nodes)
- 4 CPU
- 8GB RAM
- 100GB disk
**lib-odwork1-4** (worker nodes)
- 4 CPU
- 24GB RAM
- 100GB disk
**32 TB NFS Volume** (high capacity asset/derivative storage) | infrastructure | setup on premises cluster during we need to build out the new on premises staging environment at osu setup control plane and worker vms setup kubernetes cluster integrate with osulp rancher update ci cd build deploy pipeline provision storage test in new staging environment lib control plane nodes cpu ram disk lib worker nodes cpu ram disk tb nfs volume high capacity asset derivative storage | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.