Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
12,030 | 14,738,593,766 | IssuesEvent | 2021-01-07 05:12:23 | kdjstudios/SABillingGitlab | https://api.github.com/repos/kdjstudios/SABillingGitlab | closed | Terminated Accounts Generating Holiday Charge | anc-ops anc-process anp-important ant-bug ant-support | In GitLab by @kdjstudios on Jun 25, 2018, 15:23
**Submitted by:** "Tobey McInally" <tobey.mcinally@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-06-25-26343/conversation
**Server:** Internal (Both)
**Client/Site:** Allentown
**Account:** 4411
**Issue:**
I noticed in Allentown’s billing that our terminated accounts have drafted holiday charges (061-4411).
I understand some changes have been made to generate usage for terminated accounts, is this why the holiday fee was charged as well?
What code do you suggest will be best to use to balance these numbers? as a credit will have to be issued to get the account to a zero balance. | 1.0 | Terminated Accounts Generating Holiday Charge - In GitLab by @kdjstudios on Jun 25, 2018, 15:23
**Submitted by:** "Tobey McInally" <tobey.mcinally@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-06-25-26343/conversation
**Server:** Internal (Both)
**Client/Site:** Allentown
**Account:** 4411
**Issue:**
I noticed in Allentown’s billing that our terminated accounts have drafted holiday charges (061-4411).
I understand some changes have been made to generate usage for terminated accounts, is this why the holiday fee was charged as well?
What code do you suggest will be best to use to balance these numbers? as a credit will have to be issued to get the account to a zero balance. | process | terminated accounts generating holiday charge in gitlab by kdjstudios on jun submitted by tobey mcinally helpdesk server internal both client site allentown account issue i noticed in allentown’s billing that our terminated accounts have drafted holiday charges i understand some changes have been made to generate usage for terminated accounts is this why the holiday fee was charged as well what code do you suggest will be best to use to balance these numbers as a credit will have to be issued to get the account to a zero balance | 1 |
253,256 | 21,671,684,103 | IssuesEvent | 2022-05-08 03:24:51 | ossf/scorecard-action | https://api.github.com/repos/ossf/scorecard-action | opened | Failed to run e2e test-organization-ls/scorecard-action-private-repo-tests | e2e automated-tests | Repo: https://github.com/test-organization-ls/scorecard-action-private-repo-tests/tree/main \n Run: https://github.com/test-organization-ls/scorecard-action-private-repo-tests/actions/runs/2288360610 \n Workflow name: Scorecards-golang \n Workflow file: https://github.com/test-organization-ls/scorecard-action-private-repo-tests/tree/main/.github/workflows/Scorecards-golang.yml \n Trigger: schedule \n Branch: main \n Date: Sun May 8 03:24:36 UTC 2022 | 1.0 | Failed to run e2e test-organization-ls/scorecard-action-private-repo-tests - Repo: https://github.com/test-organization-ls/scorecard-action-private-repo-tests/tree/main \n Run: https://github.com/test-organization-ls/scorecard-action-private-repo-tests/actions/runs/2288360610 \n Workflow name: Scorecards-golang \n Workflow file: https://github.com/test-organization-ls/scorecard-action-private-repo-tests/tree/main/.github/workflows/Scorecards-golang.yml \n Trigger: schedule \n Branch: main \n Date: Sun May 8 03:24:36 UTC 2022 | non_process | failed to run test organization ls scorecard action private repo tests repo n run n workflow name scorecards golang n workflow file n trigger schedule n branch main n date sun may utc | 0 |
522,182 | 15,158,119,009 | IssuesEvent | 2021-02-12 00:22:35 | NOAA-GSL/MATS | https://api.github.com/repos/NOAA-GSL/MATS | closed | The python JSON encoder can't handle NaNs | Priority: Medium Project: MATS Status: Closed Type: Bug | ---
Author Name: **molly.b.smith** (@mollybsmith-noaa)
Original Redmine Issue: 60863, https://vlab.ncep.noaa.gov/redmine/issues/60863
Original Date: 2019-03-01
Original Assignee: molly.b.smith
---
Python's JSON module dies if you ask it to JSONify a numpy NaN, for some reason, which kills the whole python query script in METexpress.
| 1.0 | The python JSON encoder can't handle NaNs - ---
Author Name: **molly.b.smith** (@mollybsmith-noaa)
Original Redmine Issue: 60863, https://vlab.ncep.noaa.gov/redmine/issues/60863
Original Date: 2019-03-01
Original Assignee: molly.b.smith
---
Python's JSON module dies if you ask it to JSONify a numpy NaN, for some reason, which kills the whole python query script in METexpress.
| non_process | the python json encoder can t handle nans author name molly b smith mollybsmith noaa original redmine issue original date original assignee molly b smith python s json module dies if you ask it to jsonify a numpy nan for some reason which kills the whole python query script in metexpress | 0 |
34,541 | 12,292,565,931 | IssuesEvent | 2020-05-10 15:12:21 | GHPReporter/GHPReporter.github.io | https://api.github.com/repos/GHPReporter/GHPReporter.github.io | opened | CVE-2018-14042 (Medium) detected in bootstrap-3.3.6.min.js, bootstrap-3.3.6.js | security vulnerability | ## CVE-2018-14042 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.6.min.js</b>, <b>bootstrap-3.3.6.js</b></p></summary>
<p>
<details><summary><b>bootstrap-3.3.6.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /GHPReporter.github.io/dist/js/bootstrap.min.js,/GHPReporter.github.io/./dist/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-3.3.6.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /GHPReporter.github.io/dist/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/GHPReporter/GHPReporter.github.io/commit/f60db7f4f2e8c5fe12865ef80ef9a87529feab87">f60db7f4f2e8c5fe12865ef80ef9a87529feab87</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-14042 (Medium) detected in bootstrap-3.3.6.min.js, bootstrap-3.3.6.js - ## CVE-2018-14042 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.6.min.js</b>, <b>bootstrap-3.3.6.js</b></p></summary>
<p>
<details><summary><b>bootstrap-3.3.6.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /GHPReporter.github.io/dist/js/bootstrap.min.js,/GHPReporter.github.io/./dist/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-3.3.6.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /GHPReporter.github.io/dist/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/GHPReporter/GHPReporter.github.io/commit/f60db7f4f2e8c5fe12865ef80ef9a87529feab87">f60db7f4f2e8c5fe12865ef80ef9a87529feab87</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in bootstrap min js bootstrap js cve medium severity vulnerability vulnerable libraries bootstrap min js bootstrap js bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library ghpreporter github io dist js bootstrap min js ghpreporter github io dist js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library ghpreporter github io dist js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with whitesource | 0 |
1,828 | 4,613,605,535 | IssuesEvent | 2016-09-25 03:46:23 | EBrown8534/StackExchangeStatisticsExplorer | https://api.github.com/repos/EBrown8534/StackExchangeStatisticsExplorer | closed | Add Site Comparisons | enhancement in process | Add a page where two (perhaps more) sites can be compared across all metrics side-by-side. | 1.0 | Add Site Comparisons - Add a page where two (perhaps more) sites can be compared across all metrics side-by-side. | process | add site comparisons add a page where two perhaps more sites can be compared across all metrics side by side | 1 |
9,576 | 12,530,567,632 | IssuesEvent | 2020-06-04 13:17:18 | GoogleCloudPlatform/dotnet-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples | closed | Bigtable: TestListTables is timing out in CI | api: bigtable priority: p1 type: process | After 10 minutes.
I've skipped it in #1001 but it should be looked at.
@billyjacobson assigning to you because you have commited to the Bigtable samples more recently. | 1.0 | Bigtable: TestListTables is timing out in CI - After 10 minutes.
I've skipped it in #1001 but it should be looked at.
@billyjacobson assigning to you because you have commited to the Bigtable samples more recently. | process | bigtable testlisttables is timing out in ci after minutes i ve skipped it in but it should be looked at billyjacobson assigning to you because you have commited to the bigtable samples more recently | 1 |
5,514 | 8,379,150,914 | IssuesEvent | 2018-10-06 21:44:00 | carloseduardov8/Viajato | https://api.github.com/repos/carloseduardov8/Viajato | closed | Corrigir diagrama de classes | Priority:Very High Process:Create/Update Class Diagram | Altera diagrama de classes em função do novo modelo de entidades feito no JHipster, removendo alguns elementos | 1.0 | Corrigir diagrama de classes - Altera diagrama de classes em função do novo modelo de entidades feito no JHipster, removendo alguns elementos | process | corrigir diagrama de classes altera diagrama de classes em função do novo modelo de entidades feito no jhipster removendo alguns elementos | 1 |
15,141 | 18,893,982,709 | IssuesEvent | 2021-11-15 15:57:21 | beer-garden/beer-garden | https://api.github.com/repos/beer-garden/beer-garden | closed | Remedy release process deficiencies | release process | Two specific issues were encountered while preparing the last release:
1. Beer-garden's `setup.py` has dependency versions that are out of sync with the `requirements.{txt,in}` files, with the potential to introduce difficult to assess bugs in the RPM creation.
2. The version of `wrapt` that's current on PyPI will not work with beer-garden at this time. The library version should be pinned to an older version in all the appropriate places.
The best fix may be to address these two issues in the most direct way. But it might be better to figure out a mechanism by which only one of `setup.py` or `requirements.txt` will be the authoritative source of versions and have the other automatically sync with it. | 1.0 | Remedy release process deficiencies - Two specific issues were encountered while preparing the last release:
1. Beer-garden's `setup.py` has dependency versions that are out of sync with the `requirements.{txt,in}` files, with the potential to introduce difficult to assess bugs in the RPM creation.
2. The version of `wrapt` that's current on PyPI will not work with beer-garden at this time. The library version should be pinned to an older version in all the appropriate places.
The best fix may be to address these two issues in the most direct way. But it might be better to figure out a mechanism by which only one of `setup.py` or `requirements.txt` will be the authoritative source of versions and have the other automatically sync with it. | process | remedy release process deficiencies two specific issues were encountered while preparing the last release beer garden s setup py has dependency versions that are out of sync with the requirements txt in files with the potential to introduce difficult to assess bugs in the rpm creation the version of wrapt that s current on pypi will not work with beer garden at this time the library version should be pinned to an older version in all the appropriate places the best fix may be to address these two issues in the most direct way but it might be better to figure out a mechanism by which only one of setup py or requirements txt will be the authoritative source of versions and have the other automatically sync with it | 1 |
2,058 | 4,864,884,220 | IssuesEvent | 2016-11-14 19:15:55 | Sage-Bionetworks/Genie | https://api.github.com/repos/Sage-Bionetworks/Genie | opened | patients missing sample data, wrong oncotree codes | data processing MSK pending release | MSK uploaded new files to correct. Not sure whether patients were removed or sample info added. | 1.0 | patients missing sample data, wrong oncotree codes - MSK uploaded new files to correct. Not sure whether patients were removed or sample info added. | process | patients missing sample data wrong oncotree codes msk uploaded new files to correct not sure whether patients were removed or sample info added | 1 |
251,343 | 18,947,379,958 | IssuesEvent | 2021-11-18 11:40:31 | amosproj/amos2021ws05-fin-prod-port-quick-check | https://api.github.com/repos/amosproj/amos2021ws05-fin-prod-port-quick-check | closed | Use Case Diagram | est. size: 1 type: documentation type: infrastructure real size: 1 priority: low | ## User story
1. As a developer
2. I need a use case diagram
3. So that I know the different types of users, their interactions between each other and set of actions
## Acceptance criteria
* All known users are illustrated
* All known actions are illustrated
* Their interactions are illustrated
## Definition of done
* Approved by product owner
| 1.0 | Use Case Diagram - ## User story
1. As a developer
2. I need a use case diagram
3. So that I know the different types of users, their interactions between each other and set of actions
## Acceptance criteria
* All known users are illustrated
* All known actions are illustrated
* Their interactions are illustrated
## Definition of done
* Approved by product owner
| non_process | use case diagram user story as a developer i need a use case diagram so that i know the different types of users their interactions between each other and set of actions acceptance criteria all known users are illustrated all known actions are illustrated their interactions are illustrated definition of done approved by product owner | 0 |
73,230 | 7,329,713,086 | IssuesEvent | 2018-03-05 06:49:04 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Edit Ingress page allows for namespace to be changed. | area/ui kind/bug status/resolved status/to-test version/2.0 | **Rancher versions: Server build from master - Mar 2
**Steps to Reproduce:**
Create an Ingress.
Edit Ingress .
In the Ingress edit page , user is allowed to change namespace.
This should not be allowed.
<img width="1188" alt="screen shot 2018-03-02 at 10 24 32 pm" src="https://user-images.githubusercontent.com/4266958/36931600-0c9bc0c6-1e6e-11e8-969f-d9abcdc4a303.png">
| 1.0 | Edit Ingress page allows for namespace to be changed. - **Rancher versions: Server build from master - Mar 2
**Steps to Reproduce:**
Create an Ingress.
Edit Ingress .
In the Ingress edit page , user is allowed to change namespace.
This should not be allowed.
<img width="1188" alt="screen shot 2018-03-02 at 10 24 32 pm" src="https://user-images.githubusercontent.com/4266958/36931600-0c9bc0c6-1e6e-11e8-969f-d9abcdc4a303.png">
| non_process | edit ingress page allows for namespace to be changed rancher versions server build from master mar steps to reproduce create an ingress edit ingress in the ingress edit page user is allowed to change namespace this should not be allowed img width alt screen shot at pm src | 0 |
19,587 | 25,922,221,557 | IssuesEvent | 2022-12-15 23:23:32 | microsoft/cadl | https://api.github.com/repos/microsoft/cadl | closed | Variable interpolation Cannot work When input a parameter by --arg in command | bug :pushpin: WS: Process Tools & Automation Required for DPG 1.0 | ## How to reproduce
1. Download the [cadl project](https://github.com/Azure/azure-rest-api-specs/tree/dw/cadl-bug/specification/cognitiveservices/OpenAI.Inference) I used.
2. `npm install`
3. `npx cadl compile . --emit @azure-tools/cadl-java --arg "java-repo-folder=/tmp/java"`
## Result
The codes are generated in `{cwd}/azure-sdk-for-java`.
## Expected
The codes should be generated in `/tmp/java`.
## Other Issue:
In the `cadl-project.yaml`, you can find the value of `emitter-output-dir` is surrounded by double quotes. If I delete the double quotes, the cadl reports error: `bad indentation of a mapping entry (6:45)`
| 1.0 | Variable interpolation Cannot work When input a parameter by --arg in command - ## How to reproduce
1. Download the [cadl project](https://github.com/Azure/azure-rest-api-specs/tree/dw/cadl-bug/specification/cognitiveservices/OpenAI.Inference) I used.
2. `npm install`
3. `npx cadl compile . --emit @azure-tools/cadl-java --arg "java-repo-folder=/tmp/java"`
## Result
The codes are generated in `{cwd}/azure-sdk-for-java`.
## Expected
The codes should be generated in `/tmp/java`.
## Other Issue:
In the `cadl-project.yaml`, you can find the value of `emitter-output-dir` is surrounded by double quotes. If I delete the double quotes, the cadl reports error: `bad indentation of a mapping entry (6:45)`
| process | variable interpolation cannot work when input a parameter by arg in command how to reproduce download the i used npm install npx cadl compile emit azure tools cadl java arg java repo folder tmp java result the codes are generated in cwd azure sdk for java expected the codes should be generated in tmp java other issue in the cadl project yaml you can find the value of emitter output dir is surrounded by double quotes if i delete the double quotes the cadl reports error bad indentation of a mapping entry | 1 |
13,268 | 15,732,093,895 | IssuesEvent | 2021-03-29 17:52:32 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | opened | parentage: `chromosome segregation` | cell cycle and DNA processes | It seems that `chromosome separation (GO:0051304)` has these SubClass Of relationships:
* is_a `cell cycle process`
* 'part of' some 'chromosome segregation'
While `chromosome segregation` only has this SubClass Of relationship:
* is_a `cellular process`
I would think that `chromosome segregation` should also be is_a `cell cycle process`
@ValWood - any thoughts here? | 1.0 | parentage: `chromosome segregation` - It seems that `chromosome separation (GO:0051304)` has these SubClass Of relationships:
* is_a `cell cycle process`
* 'part of' some 'chromosome segregation'
While `chromosome segregation` only has this SubClass Of relationship:
* is_a `cellular process`
I would think that `chromosome segregation` should also be is_a `cell cycle process`
@ValWood - any thoughts here? | process | parentage chromosome segregation it seems that chromosome separation go has these subclass of relationships is a cell cycle process part of some chromosome segregation while chromosome segregation only has this subclass of relationship is a cellular process i would think that chromosome segregation should also be is a cell cycle process valwood any thoughts here | 1 |
8,989 | 12,100,880,405 | IssuesEvent | 2020-04-20 14:26:09 | prisma/prisma | https://api.github.com/repos/prisma/prisma | opened | SourceValidationError should be pretty printed | bug/2-confirmed kind/bug process/candidate team/engines |
## Bug description
```
Error: Error: Error in datamodel: ErrorCollection { errors: [SourceValidationError { message: "The URL for datasource `db` must start with the protocol `sqlite://`.", source: "db", span: Span { start: 55, end: 73 } }] }
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
prisma/schema.prisma file
```prisma
datasource db {
provider = "sqlite"
url = "sqlite:./1093.db"
}
```
Run `npx @prisma/cli@alpha introspect`
Enjoy!
## Expected behavior
The error should be pretty printed with a cursor pointing to the line.
## Prisma information
<!-- Your Prisma schema, Prisma Client queries, ...
Do not include your database credentials when sharing your Prisma schema! -->
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: Mac OS
- Database: SQLite
- Prisma version: 2.0.0-alpha.1108 (latest) | 1.0 | SourceValidationError should be pretty printed -
## Bug description
```
Error: Error: Error in datamodel: ErrorCollection { errors: [SourceValidationError { message: "The URL for datasource `db` must start with the protocol `sqlite://`.", source: "db", span: Span { start: 55, end: 73 } }] }
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
prisma/schema.prisma file
```prisma
datasource db {
provider = "sqlite"
url = "sqlite:./1093.db"
}
```
Run `npx @prisma/cli@alpha introspect`
Enjoy!
## Expected behavior
The error should be pretty printed with a cursor pointing to the line.
## Prisma information
<!-- Your Prisma schema, Prisma Client queries, ...
Do not include your database credentials when sharing your Prisma schema! -->
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: Mac OS
- Database: SQLite
- Prisma version: 2.0.0-alpha.1108 (latest) | process | sourcevalidationerror should be pretty printed bug description error error error in datamodel errorcollection errors how to reproduce steps to reproduce the behavior go to change run see error prisma schema prisma file prisma datasource db provider sqlite url sqlite db run npx prisma cli alpha introspect enjoy expected behavior the error should be pretty printed with a cursor pointing to the line prisma information your prisma schema prisma client queries do not include your database credentials when sharing your prisma schema environment setup os mac os database sqlite prisma version alpha latest | 1 |
213,502 | 24,003,555,243 | IssuesEvent | 2022-09-14 13:19:00 | MatBenfield/news | https://api.github.com/repos/MatBenfield/news | closed | [SecurityWeek] Today: 2022 CISO Forum Virtual Event | SecurityWeek Stale |

[read more](https://www.securityweek.com/today-2022-ciso-forum-virtual-event)
<https://www.securityweek.com/today-2022-ciso-forum-virtual-event>
| True | [SecurityWeek] Today: 2022 CISO Forum Virtual Event -

[read more](https://www.securityweek.com/today-2022-ciso-forum-virtual-event)
<https://www.securityweek.com/today-2022-ciso-forum-virtual-event>
| non_process | today ciso forum virtual event sites default files features ciso forum header jpg | 0 |
22,181 | 30,732,505,871 | IssuesEvent | 2023-07-28 03:47:57 | h4sh5/pypi-auto-scanner | https://api.github.com/repos/h4sh5/pypi-auto-scanner | opened | roblox-pyc 1.25.110 has 2 GuardDog issues | guarddog silent-process-execution | https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.25.110",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.25.110/robloxpyc/installationmanager.py:18",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.25.110/robloxpyc/installationmanager.py:25",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp05_g_zhe/roblox-pyc"
}
}``` | 1.0 | roblox-pyc 1.25.110 has 2 GuardDog issues - https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.25.110",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.25.110/robloxpyc/installationmanager.py:18",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.25.110/robloxpyc/installationmanager.py:25",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp05_g_zhe/roblox-pyc"
}
}``` | process | roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc robloxpyc installationmanager py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc installationmanager py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp g zhe roblox pyc | 1 |
264,756 | 23,137,192,412 | IssuesEvent | 2022-07-28 15:09:33 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | pkg/sql/sqlitelogictest/tests/local/local_test: TestSqlLiteLogic_testindexview10slt_good_5_test failed | C-test-failure O-robot branch-master | pkg/sql/sqlitelogictest/tests/local/local_test.TestSqlLiteLogic_testindexview10slt_good_5_test [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/5889610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/5889610?buildTab=artifacts#/) on master @ [7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732](https://github.com/cockroachdb/cockroach/commits/7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732):
```
=== RUN TestSqlLiteLogic_testindexview10slt_good_5_test
test_log_scope.go:162: test logs captured to: /artifacts/tmp/_tmp/89e8a97b552c04736e36c410c947cf4c/logTestSqlLiteLogic_testindexview10slt_good_5_test1952932785
test_log_scope.go:80: use -show-logs to present logs inline
logic.go:1725:
Error Trace: logic.go:1725
logic.go:1801
logic.go:3657
generated_test.go:74
generated_test.go:1320
Error: Received unexpected error:
open /test/index/view/10/slt_good_5.test: no such file or directory
Test: TestSqlLiteLogic_testindexview10slt_good_5_test
panic.go:500: -- test log scope end --
--- FAIL: TestSqlLiteLogic_testindexview10slt_good_5_test (0.02s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSqlLiteLogic_testindexview10slt_good_5_test.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-18093 | 1.0 | pkg/sql/sqlitelogictest/tests/local/local_test: TestSqlLiteLogic_testindexview10slt_good_5_test failed - pkg/sql/sqlitelogictest/tests/local/local_test.TestSqlLiteLogic_testindexview10slt_good_5_test [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/5889610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/5889610?buildTab=artifacts#/) on master @ [7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732](https://github.com/cockroachdb/cockroach/commits/7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732):
```
=== RUN TestSqlLiteLogic_testindexview10slt_good_5_test
test_log_scope.go:162: test logs captured to: /artifacts/tmp/_tmp/89e8a97b552c04736e36c410c947cf4c/logTestSqlLiteLogic_testindexview10slt_good_5_test1952932785
test_log_scope.go:80: use -show-logs to present logs inline
logic.go:1725:
Error Trace: logic.go:1725
logic.go:1801
logic.go:3657
generated_test.go:74
generated_test.go:1320
Error: Received unexpected error:
open /test/index/view/10/slt_good_5.test: no such file or directory
Test: TestSqlLiteLogic_testindexview10slt_good_5_test
panic.go:500: -- test log scope end --
--- FAIL: TestSqlLiteLogic_testindexview10slt_good_5_test (0.02s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSqlLiteLogic_testindexview10slt_good_5_test.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-18093 | non_process | pkg sql sqlitelogictest tests local local test testsqllitelogic good test failed pkg sql sqlitelogictest tests local local test testsqllitelogic good test with on master run testsqllitelogic good test test log scope go test logs captured to artifacts tmp tmp logtestsqllitelogic good test log scope go use show logs to present logs inline logic go error trace logic go logic go logic go generated test go generated test go error received unexpected error open test index view slt good test no such file or directory test testsqllitelogic good test panic go test log scope end fail testsqllitelogic good test help see also jira issue crdb | 0 |
9,455 | 12,438,295,840 | IssuesEvent | 2020-05-26 08:10:15 | aiidateam/aiida-core | https://api.github.com/repos/aiidateam/aiida-core | closed | Add support for type hint parsing in automatic process spec determination for process functions | priority/nice-to-have topic/processes type/accepted feature type/duplicate | Now that we support only Python versions that have type hinting for function signatures, we could maybe inspect those hints to improve the automatically generated process spec to set the `valid_type` based on it. This will improve for example the output of `verdi plugin list` with the description of process functions which now will always just show `Data` as the required type. | 1.0 | Add support for type hint parsing in automatic process spec determination for process functions - Now that we support only Python versions that have type hinting for function signatures, we could maybe inspect those hints to improve the automatically generated process spec to set the `valid_type` based on it. This will improve for example the output of `verdi plugin list` with the description of process functions which now will always just show `Data` as the required type. | process | add support for type hint parsing in automatic process spec determination for process functions now that we support only python versions that have type hinting for function signatures we could maybe inspect those hints to improve the automatically generated process spec to set the valid type based on it this will improve for example the output of verdi plugin list with the description of process functions which now will always just show data as the required type | 1 |
10,677 | 13,462,434,851 | IssuesEvent | 2020-09-09 16:04:03 | tdwg/dwc | https://api.github.com/repos/tdwg/dwc | closed | DD-MM-YYYY and MM-DD-YYYY in eventDate | Class - Event Docs - Text Guide Format - Text Process - dismissed answered | dwc guidance documents should explicitly recommend against these commonly human-used forms since it is impossible for some dates to distinguish between the U.S. usage MM-DD-YYYY and the European usage DD-MM-YYYY. A casual misreading of ISO8601 might lead one to (erroneously) believe that these are compliant with ISO8601 and therefore compliant with dwc:eventDate suggestions.
I've seen these forms in published DwC-A in the wild.
| 1.0 | DD-MM-YYYY and MM-DD-YYYY in eventDate - dwc guidance documents should explicitly recommend against these commonly human-used forms since it is impossible for some dates to distinguish between the U.S. usage MM-DD-YYYY and the European usage DD-MM-YYYY. A casual misreading of ISO8601 might lead one to (erroneously) believe that these are compliant with ISO8601 and therefore compliant with dwc:eventDate suggestions.
I've seen these forms in published DwC-A in the wild.
| process | dd mm yyyy and mm dd yyyy in eventdate dwc guidance documents should explicitly recommend against these commonly human used forms since it is impossible for some dates to distinguish between the u s usage mm dd yyyy and the european usage dd mm yyyy a casual misreading of might lead one to erroneously believe that these are compliant with and therefore compliant with dwc eventdate suggestions i ve seen these forms in published dwc a in the wild | 1 |
214,008 | 16,544,379,250 | IssuesEvent | 2021-05-27 21:25:02 | xuhanz/DailyDose | https://api.github.com/repos/xuhanz/DailyDose | opened | Developer Guideline Issues | documentation | Add more instructions regarding the format of newly added tests so they can fit in with existing tests. | 1.0 | Developer Guideline Issues - Add more instructions regarding the format of newly added tests so they can fit in with existing tests. | non_process | developer guideline issues add more instructions regarding the format of newly added tests so they can fit in with existing tests | 0 |
291,971 | 21,945,031,403 | IssuesEvent | 2022-05-23 22:56:21 | getditto/docs | https://api.github.com/repos/getditto/docs | closed | Problem: Missing write transaction section in concepts | documentation | # Problem
Ditto has long had support for write transactions across collections. However the documentation doesn't have a `Concepts > Write Transaction` section
# Solution
Please create a section called `Concepts > Write Transactions` and include necessary snippets. | 1.0 | Problem: Missing write transaction section in concepts - # Problem
Ditto has long had support for write transactions across collections. However the documentation doesn't have a `Concepts > Write Transaction` section
# Solution
Please create a section called `Concepts > Write Transactions` and include necessary snippets. | non_process | problem missing write transaction section in concepts problem ditto has long had support for write transactions across collections however the documentation doesn t have a concepts write transaction section solution please create a section called concepts write transactions and include necessary snippets | 0 |
11,496 | 14,368,809,579 | IssuesEvent | 2020-12-01 08:56:13 | panther-labs/panther | https://api.github.com/repos/panther-labs/panther | closed | Introduce Custom Logs apis | p1 story team:data processing | ### Description
Introduce the BE components for supporting Custom Logs
### Acceptance Criteria
All Custom Logs-related code is in OSS. | 1.0 | Introduce Custom Logs apis - ### Description
Introduce the BE components for supporting Custom Logs
### Acceptance Criteria
All Custom Logs-related code is in OSS. | process | introduce custom logs apis description introduce the be components for supporting custom logs acceptance criteria all custom logs related code is in oss | 1 |
11,708 | 14,545,565,701 | IssuesEvent | 2020-12-15 19:50:56 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | closed | multi-job output variable named Foo.Bar | Pri1 devops-cicd-process/tech devops/prod doc-enhancement stale-issue | It is explained in this section how to grab a variable output from one job, in another job: https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops&tabs=yaml%2Cbatch#set-a-multi-job-output-variable
However the variable name is simply "myOutputVar" in your example, and you access it like this:
```
myVarFromJobA: $[ dependencies.A.outputs['setvarStep.myOutputVar'] ]
```
However in my case, the variable I need to access is set by an executable and it's named "GitVersion.SemVer".
How do I access that using this syntax? I tried this, but the extra "." in the property name doesn't seem to work:
myVarFromJobA: $[ dependencies.A.outputs['setvarStep.GitVersion.SemVer'] ]
I have tried to guess other ways and they all failed:
myVarFromJobA: $[ dependencies.A.outputs['setvarStep.variables['GitVersion.SemVer']]' ]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a
* Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a
* Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops&tabs=yaml%2Cbatch#set-a-multi-job-output-variable)
* Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | multi-job output variable named Foo.Bar - It is explained in this section how to grab a variable output from one job, in another job: https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops&tabs=yaml%2Cbatch#set-a-multi-job-output-variable
However the variable name is simply "myOutputVar" in your example, and you access it like this:
```
myVarFromJobA: $[ dependencies.A.outputs['setvarStep.myOutputVar'] ]
```
However in my case, the variable I need to access is set by an executable and it's named "GitVersion.SemVer".
How do I access that using this syntax? I tried this, but the extra "." in the property name doesn't seem to work:
myVarFromJobA: $[ dependencies.A.outputs['setvarStep.GitVersion.SemVer'] ]
I have tried to guess other ways and they all failed:
myVarFromJobA: $[ dependencies.A.outputs['setvarStep.variables['GitVersion.SemVer']]' ]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a
* Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a
* Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops&tabs=yaml%2Cbatch#set-a-multi-job-output-variable)
* Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | process | multi job output variable named foo bar it is explained in this section how to grab a variable output from one job in another job however the variable name is simply myoutputvar in your example and you access it like this myvarfromjoba however in my case the variable i need to access is set by an executable and it s named gitversion semver how do i access that using this syntax i tried this but the extra in the property name doesn t seem to work myvarfromjoba i have tried to guess other ways and they all failed myvarfromjoba document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id bcdb content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam | 1 |
2,522 | 5,287,965,563 | IssuesEvent | 2017-02-08 13:58:16 | Hurence/logisland | https://api.github.com/repos/Hurence/logisland | opened | add possibility to use setRouting in putElasticSearch processor | enhancement processor | You can use this to specify a string that will be used to gather document having the same routing in the same shard in order to optimize perfs. | 1.0 | add possibility to use setRouting in putElasticSearch processor - You can use this to specify a string that will be used to gather document having the same routing in the same shard in order to optimize perfs. | process | add possibility to use setrouting in putelasticsearch processor you can use this to specify a string that will be used to gather document having the same routing in the same shard in order to optimize perfs | 1 |
19,669 | 26,029,287,789 | IssuesEvent | 2022-12-21 19:23:53 | googleapis/google-cloud-go | https://api.github.com/repos/googleapis/google-cloud-go | closed | all: broken genbot CI build | type: process | The `gapicgen/genbot` build is currently broken ([log](https://fusion2.corp.google.com/invocations/43586277-80b3-42da-8e5d-96c85d496309/targets/cloud-devrel%2Fclient-libraries%2Fgo%2Fgapicgen%2Fgenbot/log)) and may continue to encounter the type of error shown below until we have dropped support for older versions of Go (< 1.17 ?). Until then, we should consider including the go version flags suggested below in our builds as needed.
```
[cloud.google.com/go/accessapproval/apiv1](http://cloud.google.com/go/accessapproval/apiv1) imports
[google.golang.org/grpc/codes](https://www.google.com/url?q=http://google.golang.org/grpc/codes&sa=D) tested by
[google.golang.org/grpc/codes.test](https://www.google.com/url?q=http://google.golang.org/grpc/codes.test&sa=D) imports
[google.golang.org/genproto/googleapis/rpc/code](https://www.google.com/url?q=http://google.golang.org/genproto/googleapis/rpc/code&sa=D) loaded from [google.golang.org/genproto@v0.0.0-20220920201722-2b89144ce006](https://www.google.com/url?q=http://google.golang.org/genproto@v0.0.0-20220920201722-2b89144ce006&sa=D),
but go 1.16 would select v0.0.0-20220926220553-6981cbe3cfce
To upgrade to the versions selected by go 1.16:
go mod tidy -go=1.16 && go mod tidy -go=1.17
If reproducibility with go 1.16 is not needed:
go mod tidy -compat=1.17
For other options, see:
[https://golang.org/doc/modules/pruning](https://www.google.com/url?q=https://golang.org/doc/modules/pruning&sa=D)
error generating gapics (may need to check logs for more errors): exit status 1
exit status 1
``` | 1.0 | all: broken genbot CI build - The `gapicgen/genbot` build is currently broken ([log](https://fusion2.corp.google.com/invocations/43586277-80b3-42da-8e5d-96c85d496309/targets/cloud-devrel%2Fclient-libraries%2Fgo%2Fgapicgen%2Fgenbot/log)) and may continue to encounter the type of error shown below until we have dropped support for older versions of Go (< 1.17 ?). Until then, we should consider including the go version flags suggested below in our builds as needed.
```
[cloud.google.com/go/accessapproval/apiv1](http://cloud.google.com/go/accessapproval/apiv1) imports
[google.golang.org/grpc/codes](https://www.google.com/url?q=http://google.golang.org/grpc/codes&sa=D) tested by
[google.golang.org/grpc/codes.test](https://www.google.com/url?q=http://google.golang.org/grpc/codes.test&sa=D) imports
[google.golang.org/genproto/googleapis/rpc/code](https://www.google.com/url?q=http://google.golang.org/genproto/googleapis/rpc/code&sa=D) loaded from [google.golang.org/genproto@v0.0.0-20220920201722-2b89144ce006](https://www.google.com/url?q=http://google.golang.org/genproto@v0.0.0-20220920201722-2b89144ce006&sa=D),
but go 1.16 would select v0.0.0-20220926220553-6981cbe3cfce
To upgrade to the versions selected by go 1.16:
go mod tidy -go=1.16 && go mod tidy -go=1.17
If reproducibility with go 1.16 is not needed:
go mod tidy -compat=1.17
For other options, see:
[https://golang.org/doc/modules/pruning](https://www.google.com/url?q=https://golang.org/doc/modules/pruning&sa=D)
error generating gapics (may need to check logs for more errors): exit status 1
exit status 1
``` | process | all broken genbot ci build the gapicgen genbot build is currently broken and may continue to encounter the type of error shown below until we have dropped support for older versions of go until then we should consider including the go version flags suggested below in our builds as needed imports tested by imports loaded from but go would select to upgrade to the versions selected by go go mod tidy go go mod tidy go if reproducibility with go is not needed go mod tidy compat for other options see error generating gapics may need to check logs for more errors exit status exit status | 1 |
6,386 | 9,460,102,452 | IssuesEvent | 2019-04-17 10:05:16 | googleapis/google-cloud-dotnet | https://api.github.com/repos/googleapis/google-cloud-dotnet | closed | Consider making Spanner libraries v2 only target netstandard2.0 | api: spanner type: process | This is similar to #2959, but for Spanner.
Currently we have the following targeting:
- Spanner.Data: netstandard1.5;netstandard2.0;net45
- Spanner.Common.V1: netstandard1.5;net45
- Spanner.V1: netstandard1.5;netstandard2.0;net45
- Spanner.Admin.Database.V1: netstandard1.5;net45
- Spanner.Admin.Instance.V1: netstandard1.5;net45
Proposal: make all of these *just* target netstandard2.0.
This makes it simpler to reason about and simpler to test.
cc for input:
@FransBouma
@tomerpeled
@alyaros
@jerarl
@SurferJeffAtGoogle
@chrisdunelm
@amanda-tarafa
We may well want to change to make this the target for anything that hasn't already gone GA... | 1.0 | Consider making Spanner libraries v2 only target netstandard2.0 - This is similar to #2959, but for Spanner.
Currently we have the following targeting:
- Spanner.Data: netstandard1.5;netstandard2.0;net45
- Spanner.Common.V1: netstandard1.5;net45
- Spanner.V1: netstandard1.5;netstandard2.0;net45
- Spanner.Admin.Database.V1: netstandard1.5;net45
- Spanner.Admin.Instance.V1: netstandard1.5;net45
Proposal: make all of these *just* target netstandard2.0.
This makes it simpler to reason about and simpler to test.
cc for input:
@FransBouma
@tomerpeled
@alyaros
@jerarl
@SurferJeffAtGoogle
@chrisdunelm
@amanda-tarafa
We may well want to change to make this the target for anything that hasn't already gone GA... | process | consider making spanner libraries only target this is similar to but for spanner currently we have the following targeting spanner data spanner common spanner spanner admin database spanner admin instance proposal make all of these just target this makes it simpler to reason about and simpler to test cc for input fransbouma tomerpeled alyaros jerarl surferjeffatgoogle chrisdunelm amanda tarafa we may well want to change to make this the target for anything that hasn t already gone ga | 1 |
13,214 | 15,685,888,834 | IssuesEvent | 2021-03-25 11:48:59 | threefoldtech/js-sdk | https://api.github.com/repos/threefoldtech/js-sdk | closed | Extension of storage nodes systematically leads to timeouts | process_wontfix type_bug | I extended my VDC with storage capacity. I see the amount taken away from my balance quite immediately, but the VDC payment screen keeps on spinning, and gets a timeout. I had this a few times (and no successful try), so I imagine there is an issue.

In TF Connect app, money is returned (apart from the fee).
Network: testnet
VDC: geertincubaid - vdcdiamondtest2 | 1.0 | Extension of storage nodes systematically leads to timeouts - I extended my VDC with storage capacity. I see the amount taken away from my balance quite immediately, but the VDC payment screen keeps on spinning, and gets a timeout. I had this a few times (and no successful try), so I imagine there is an issue.

In TF Connect app, money is returned (apart from the fee).
Network: testnet
VDC: geertincubaid - vdcdiamondtest2 | process | extension of storage nodes systematically leads to timeouts i extended my vdc with storage capacity i see the amount taken away from my balance quite immediately but the vdc payment screen keeps on spinning and gets a timeout i had this a few times and no successful try so i imagine there is an issue in tf connect app money is returned apart from the fee network testnet vdc geertincubaid | 1 |
167,171 | 6,333,515,288 | IssuesEvent | 2017-07-26 14:53:21 | oSoc17/oasis-frontend | https://api.github.com/repos/oSoc17/oasis-frontend | closed | Publication on oasis.team | Priority 1 | Write a publication for oasis.team which describes our project and what we've done. | 1.0 | Publication on oasis.team - Write a publication for oasis.team which describes our project and what we've done. | non_process | publication on oasis team write a publication for oasis team which describes our project and what we ve done | 0 |
4,534 | 7,373,159,708 | IssuesEvent | 2018-03-13 16:31:26 | Activiti/Activiti | https://api.github.com/repos/Activiti/Activiti | opened | The process instance API doesn't return the parent information | blocking process | Even though the process definition contains a subprocess, the API doesn't return the parent information.
It seems from the server side the methods getParentId() and getRootProcessInstanceId() are available, but those are missing at rest api level.
That information should be returned by the API
http://{{domain}}/query/v1/process-instances
and runtime bundle too
http://{{domain}}/rb-my-app/v1/process-instances | 1.0 | The process instance API doesn't return the parent information - Even though the process definition contains a subprocess, the API doesn't return the parent information.
It seems from the server side the methods getParentId() and getRootProcessInstanceId() are available, but those are missing at rest api level.
That information should be returned by the API
http://{{domain}}/query/v1/process-instances
and runtime bundle too
http://{{domain}}/rb-my-app/v1/process-instances | process | the process instance api doesn t return the parent information even though the process definition contains a subprocess the api doesn t return the parent information it seems from the server side the methods getparentid and getrootprocessinstanceid are available but those are missing at rest api level that information should be returned by the api and runtime bundle too | 1 |
172,636 | 14,370,412,830 | IssuesEvent | 2020-12-01 11:03:56 | xmos/lib_logging | https://api.github.com/repos/xmos/lib_logging | opened | examples/AN00239 Application Note not compliant | type:documentation type:enhancement | **Issues:**
- xdoc-custom RST directives used in the Application Note README
- application note code not running successfully on xcore-ai
**Application note README**
This file contains the `appdeps`, `appnote`, and `version` directives.
**Not running on xcore-ai**
Changing the target in the Makefile to `TARGET = XCORE-AI-EXPLORER` allows the Application Note executable to run successfully on an xcore-ai. The Makefile requires enhancement to support both xcore-200 and xcore-ai.
| 1.0 | examples/AN00239 Application Note not compliant - **Issues:**
- xdoc-custom RST directives used in the Application Note README
- application note code not running successfully on xcore-ai
**Application note README**
This file contains the `appdeps`, `appnote`, and `version` directives.
**Not running on xcore-ai**
Changing the target in the Makefile to `TARGET = XCORE-AI-EXPLORER` allows the Application Note executable to run successfully on an xcore-ai. The Makefile requires enhancement to support both xcore-200 and xcore-ai.
| non_process | examples application note not compliant issues xdoc custom rst directives used in the application note readme application note code not running successfully on xcore ai application note readme this file contains the appdeps appnote and version directives not running on xcore ai changing the target in the makefile to target xcore ai explorer allows the application note executable to run successfully on an xcore ai the makefile requires enhancement to support both xcore and xcore ai | 0 |
1,557 | 4,159,632,045 | IssuesEvent | 2016-06-17 09:50:28 | openvstorage/volumedriver | https://api.github.com/repos/openvstorage/volumedriver | closed | Remove rollback API from python client | priority_minor process_wontfix type_feature | While discussing openvstorage/framework#540 it turned out that doing a rollback can cause issues. We should in that case also clean up that API on the python client. | 1.0 | Remove rollback API from python client - While discussing openvstorage/framework#540 it turned out that doing a rollback can cause issues. We should in that case also clean up that API on the python client. | process | remove rollback api from python client while discussing openvstorage framework it turned out that doing a rollback can cause issues we should in that case also clean up that api on the python client | 1 |
206,206 | 23,367,792,157 | IssuesEvent | 2022-08-10 16:52:25 | turkdevops/next-auth-example | https://api.github.com/repos/turkdevops/next-auth-example | closed | CVE-2022-31186 (Low) detected in next-auth-3.29.8.tgz - autoclosed | security vulnerability | ## CVE-2022-31186 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>next-auth-3.29.8.tgz</b></p></summary>
<p>Authentication for Next.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/next-auth/-/next-auth-3.29.8.tgz">https://registry.npmjs.org/next-auth/-/next-auth-3.29.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/next-auth/package.json</p>
<p>
Dependency Hierarchy:
- :x: **next-auth-3.29.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/next-auth-example/commit/1aaa89a68971e6811042f781fce5f59a4a7ae169">1aaa89a68971e6811042f781fce5f59a4a7ae169</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
NextAuth.js is a complete open source authentication solution for Next.js applications. An information disclosure vulnerability in `next-auth` before `v4.10.2` and `v3.29.9` allows an attacker with log access privilege to obtain excessive information such as an identity provider's secret in the log (which is thrown during OAuth error handling) and use it to leverage further attacks on the system, like impersonating the client to ask for extensive permissions. This issue has been patched in `v4.10.2` and `v3.29.9` by moving the log for `provider` information to the debug level. In addition, we added a warning for having the `debug: true` option turned on in production. If for some reason you cannot upgrade, you can user the `logger` configuration option by sanitizing the logs.
<p>Publish Date: 2022-08-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31186>CVE-2022-31186</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/nextauthjs/next-auth/security/advisories/GHSA-p6mm-27gq-9v3p">https://github.com/nextauthjs/next-auth/security/advisories/GHSA-p6mm-27gq-9v3p</a></p>
<p>Release Date: 2022-08-01</p>
<p>Fix Resolution: next-auth - 3.29.9,4.10.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-31186 (Low) detected in next-auth-3.29.8.tgz - autoclosed - ## CVE-2022-31186 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>next-auth-3.29.8.tgz</b></p></summary>
<p>Authentication for Next.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/next-auth/-/next-auth-3.29.8.tgz">https://registry.npmjs.org/next-auth/-/next-auth-3.29.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/next-auth/package.json</p>
<p>
Dependency Hierarchy:
- :x: **next-auth-3.29.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/next-auth-example/commit/1aaa89a68971e6811042f781fce5f59a4a7ae169">1aaa89a68971e6811042f781fce5f59a4a7ae169</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
NextAuth.js is a complete open source authentication solution for Next.js applications. An information disclosure vulnerability in `next-auth` before `v4.10.2` and `v3.29.9` allows an attacker with log access privilege to obtain excessive information such as an identity provider's secret in the log (which is thrown during OAuth error handling) and use it to leverage further attacks on the system, like impersonating the client to ask for extensive permissions. This issue has been patched in `v4.10.2` and `v3.29.9` by moving the log for `provider` information to the debug level. In addition, we added a warning for having the `debug: true` option turned on in production. If for some reason you cannot upgrade, you can user the `logger` configuration option by sanitizing the logs.
<p>Publish Date: 2022-08-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31186>CVE-2022-31186</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/nextauthjs/next-auth/security/advisories/GHSA-p6mm-27gq-9v3p">https://github.com/nextauthjs/next-auth/security/advisories/GHSA-p6mm-27gq-9v3p</a></p>
<p>Release Date: 2022-08-01</p>
<p>Fix Resolution: next-auth - 3.29.9,4.10.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve low detected in next auth tgz autoclosed cve low severity vulnerability vulnerable library next auth tgz authentication for next js library home page a href path to dependency file package json path to vulnerable library node modules next auth package json dependency hierarchy x next auth tgz vulnerable library found in head commit a href found in base branch main vulnerability details nextauth js is a complete open source authentication solution for next js applications an information disclosure vulnerability in next auth before and allows an attacker with log access privilege to obtain excessive information such as an identity provider s secret in the log which is thrown during oauth error handling and use it to leverage further attacks on the system like impersonating the client to ask for extensive permissions this issue has been patched in and by moving the log for provider information to the debug level in addition we added a warning for having the debug true option turned on in production if for some reason you cannot upgrade you can user the logger configuration option by sanitizing the logs publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution next auth step up your open source security game with mend | 0 |
17,223 | 22,833,626,299 | IssuesEvent | 2022-07-12 14:50:02 | googleapis/google-cloud-dotnet | https://api.github.com/repos/googleapis/google-cloud-dotnet | closed | Remove the dependencies branch and everything that needs it | type: process | The dependencies branch is now quite out-of-date, and it's fiddly to update.
I believe we use it for the documentation that's still on googleapis.dev.
I seem to remember that I've tried to remove this in the past, but was unsuccessful in my attempt.
Given that the googleapis.dev documentation is to some extent "best effort" and not the canonical documentation (which is on cloud.google.com for Cloud APIs) I think it would be reasonable for us not to have all dependencies linked there.
| 1.0 | Remove the dependencies branch and everything that needs it - The dependencies branch is now quite out-of-date, and it's fiddly to update.
I believe we use it for the documentation that's still on googleapis.dev.
I seem to remember that I've tried to remove this in the past, but was unsuccessful in my attempt.
Given that the googleapis.dev documentation is to some extent "best effort" and not the canonical documentation (which is on cloud.google.com for Cloud APIs) I think it would be reasonable for us not to have all dependencies linked there.
| process | remove the dependencies branch and everything that needs it the dependencies branch is now quite out of date and it s fiddly to update i believe we use it for the documentation that s still on googleapis dev i seem to remember that i ve tried to remove this in the past but was unsuccessful in my attempt given that the googleapis dev documentation is to some extent best effort and not the canonical documentation which is on cloud google com for cloud apis i think it would be reasonable for us not to have all dependencies linked there | 1 |
22,524 | 31,625,465,192 | IssuesEvent | 2023-09-06 04:50:04 | neurobagel/planning | https://api.github.com/repos/neurobagel/planning | closed | Create manual label push workflow | type:maintenance maint:process | When I create a new repo or when there is a repo that doesn't have a large number of our otherwise synced labels, then I want to be able to launch a workflow manually that takes all the labels in this (planning) repo and pushes them to all the repos in this organization (ignoring renames, but aware of edits other than name).
https://github.com/neurobagel/planning/blob/c24cd5117e04bce5c4b32db346af995f6a451c0a/.github/workflows/manual_wf.yml is pretty close but creates too many matrix jobs so we exceed the usage limit from Github: https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration#usage-limits
We should find a way to achieve this without exceeding the limit. | 1.0 | Create manual label push workflow - When I create a new repo or when there is a repo that doesn't have a large number of our otherwise synced labels, then I want to be able to launch a workflow manually that takes all the labels in this (planning) repo and pushes them to all the repos in this organization (ignoring renames, but aware of edits other than name).
https://github.com/neurobagel/planning/blob/c24cd5117e04bce5c4b32db346af995f6a451c0a/.github/workflows/manual_wf.yml is pretty close but creates too many matrix jobs so we exceed the usage limit from Github: https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration#usage-limits
We should find a way to achieve this without exceeding the limit. | process | create manual label push workflow when i create a new repo or when there is a repo that doesn t have a large number of our otherwise synced labels then i want to be able to launch a workflow manually that takes all the labels in this planning repo and pushes them to all the repos in this organization ignoring renames but aware of edits other than name is pretty close but creates too many matrix jobs so we exceed the usage limit from github we should find a way to achieve this without exceeding the limit | 1 |
12,313 | 14,877,319,376 | IssuesEvent | 2021-01-20 02:51:58 | CATcher-org/CATcher | https://api.github.com/repos/CATcher-org/CATcher | closed | Setup automated deployment for CATcher's web app | aspect-Process | Our goal should be to deploy the app on every commit to the master branch.
This will allow us to have a test-able version of the webapp, after every change.
We can setup a gh-pages site on the CATcher main repository, for this purpose.
The `release` version of CATcher's web app can be deployed on a separate gh-pages. | 1.0 | Setup automated deployment for CATcher's web app - Our goal should be to deploy the app on every commit to the master branch.
This will allow us to have a test-able version of the webapp, after every change.
We can setup a gh-pages site on the CATcher main repository, for this purpose.
The `release` version of CATcher's web app can be deployed on a separate gh-pages. | process | setup automated deployment for catcher s web app our goal should be to deploy the app on every commit to the master branch this will allow us to have a test able version of the webapp after every change we can setup a gh pages site on the catcher main repository for this purpose the release version of catcher s web app can be deployed on a separate gh pages | 1 |
20,905 | 3,856,701,220 | IssuesEvent | 2016-04-07 00:32:43 | letsencrypt/letsencrypt | https://api.github.com/repos/letsencrypt/letsencrypt | closed | Add MariaDB debugging to Travis build | testing | In Boulder, we just landed this change to our Travis config: https://github.com/letsencrypt/boulder/pull/1625. It sends mysql.err to stdout in the event of failure. Hopefully this will give us more insight into the spurious i/o timeout failures.
Unfortunately to be able to do this, I had to set sudo:true (temporarily), which puts us on the slower, non-container infrastructure. I don't yet have an idea of how much slower.
Note that since the client gets more pull requests than Boulder, it tends to trigger the bug more often. You may want to consider making the same change in the client .travis.yml. | 1.0 | Add MariaDB debugging to Travis build - In Boulder, we just landed this change to our Travis config: https://github.com/letsencrypt/boulder/pull/1625. It sends mysql.err to stdout in the event of failure. Hopefully this will give us more insight into the spurious i/o timeout failures.
Unfortunately to be able to do this, I had to set sudo:true (temporarily), which puts us on the slower, non-container infrastructure. I don't yet have an idea of how much slower.
Note that since the client gets more pull requests than Boulder, it tends to trigger the bug more often. You may want to consider making the same change in the client .travis.yml. | non_process | add mariadb debugging to travis build in boulder we just landed this change to our travis config it sends mysql err to stdout in the event of failure hopefully this will give us more insight into the spurious i o timeout failures unfortunately to be able to do this i had to set sudo true temporarily which puts us on the slower non container infrastructure i don t yet have an idea of how much slower note that since the client gets more pull requests than boulder it tends to trigger the bug more often you may want to consider making the same change in the client travis yml | 0 |
18,888 | 24,826,045,369 | IssuesEvent | 2022-10-25 20:44:22 | maticnetwork/miden | https://api.github.com/repos/maticnetwork/miden | closed | Optimize range checker execution trace finalization / length check | enhancement processor | Currently, the 8-bit range check table is built twice when the execution trace is finalized.
1. It is built during the [call to `RangeChecker::trace_len`](https://github.com/maticnetwork/miden/blob/78f4b7d228aca6d0949ac37dc9854d6662a2ee8a/processor/src/trace.rs#L222) when the maximum length of the execution trace is being computed in `finalize_trace`.
2. It is built during the [`range.into_trace` call](https://github.com/maticnetwork/miden/blob/78f4b7d228aca6d0949ac37dc9854d6662a2ee8a/processor/src/trace.rs#L240) before the range checker's trace is combined into the final execution trace.
Ideally, it would be better to build this table just once. It is probably also worth checking the finalization of other processors to ensure there are no similar duplicated computations. | 1.0 | Optimize range checker execution trace finalization / length check - Currently, the 8-bit range check table is built twice when the execution trace is finalized.
1. It is built during the [call to `RangeChecker::trace_len`](https://github.com/maticnetwork/miden/blob/78f4b7d228aca6d0949ac37dc9854d6662a2ee8a/processor/src/trace.rs#L222) when the maximum length of the execution trace is being computed in `finalize_trace`.
2. It is built during the [`range.into_trace` call](https://github.com/maticnetwork/miden/blob/78f4b7d228aca6d0949ac37dc9854d6662a2ee8a/processor/src/trace.rs#L240) before the range checker's trace is combined into the final execution trace.
Ideally, it would be better to build this table just once. It is probably also worth checking the finalization of other processors to ensure there are no similar duplicated computations. | process | optimize range checker execution trace finalization length check currently the bit range check table is built twice when the execution trace is finalized it is built during the when the maximum length of the execution trace is being computed in finalize trace it is built during the before the range checker s trace is combined into the final execution trace ideally it would be better to build this table just once it is probably also worth checking the finalization of other processors to ensure there are no similar duplicated computations | 1 |
6,002 | 8,808,924,815 | IssuesEvent | 2018-12-27 16:55:17 | linnovate/root | https://api.github.com/repos/linnovate/root | closed | multiple selection in every entity select everything bug | 2.0.6 Fixed Process bug critical | when you press "select everything" in an entity, and try to update , it selects only the first entity in the list, and only updates that one

after opening atleast 60 entities, it only selects a few out of the list instead of all of them

| 1.0 | multiple selection in every entity select everything bug - when you press "select everything" in an entity, and try to update , it selects only the first entity in the list, and only updates that one

after opening atleast 60 entities, it only selects a few out of the list instead of all of them

| process | multiple selection in every entity select everything bug when you press select everything in an entity and try to update it selects only the first entity in the list and only updates that one after opening atleast entities it only selects a few out of the list instead of all of them | 1 |
18,497 | 24,551,078,990 | IssuesEvent | 2022-10-12 12:39:48 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [iOS] [Offline indicator] Enrollment flow > the following offline error message should get displayed when user is offline in the following scenario | Bug P1 iOS Process: Fixed Process: Tested dev | Steps:
1. Sign up or sign in to the mobile app
2. Click on the study
3. Click on the Participate
4. Turn off the data and observe
AR: Offline error message is not getting displayed
ER: 'You are offline, You can still use this section but may miss out on the latest content updates' error message should get displayed
[Note: Issue should be fixed for Open study closed study and consent update after clicking on the review button]
For error message UI please refer to the following attached screen

| 2.0 | [iOS] [Offline indicator] Enrollment flow > the following offline error message should get displayed when user is offline in the following scenario - Steps:
1. Sign up or sign in to the mobile app
2. Click on the study
3. Click on the Participate
4. Turn off the data and observe
AR: Offline error message is not getting displayed
ER: 'You are offline, You can still use this section but may miss out on the latest content updates' error message should get displayed
[Note: Issue should be fixed for Open study closed study and consent update after clicking on the review button]
For error message UI please refer to the following attached screen

| process | enrollment flow the following offline error message should get displayed when user is offline in the following scenario steps sign up or sign in to the mobile app click on the study click on the participate turn off the data and observe ar offline error message is not getting displayed er you are offline you can still use this section but may miss out on the latest content updates error message should get displayed for error message ui please refer to the following attached screen | 1 |
16,488 | 21,445,564,491 | IssuesEvent | 2022-04-25 05:45:54 | zotero/zotero | https://api.github.com/repos/zotero/zotero | opened | Classic citation dialog: Add all selected items to Multiple Sources pane | Papercuts Word Processor Integration Bug | https://forums.zotero.org/discussion/96671/citing-multiple-papers
Not that we want to do more to the classic dialog, but this seems like a trivial fix to a clear bug (albeit one that has probably existed for a decade), and it will help people until we have a new citation dialog. | 1.0 | Classic citation dialog: Add all selected items to Multiple Sources pane - https://forums.zotero.org/discussion/96671/citing-multiple-papers
Not that we want to do more to the classic dialog, but this seems like a trivial fix to a clear bug (albeit one that has probably existed for a decade), and it will help people until we have a new citation dialog. | process | classic citation dialog add all selected items to multiple sources pane not that we want to do more to the classic dialog but this seems like a trivial fix to a clear bug albeit one that has probably existed for a decade and it will help people until we have a new citation dialog | 1 |
14,993 | 18,674,461,890 | IssuesEvent | 2021-10-31 10:09:55 | googleapis/python-bigquery | https://api.github.com/repos/googleapis/python-bigquery | opened | Bump minimum openetelemetry version to support type checks | type: process | See the following comment for details: https://github.com/googleapis/python-bigquery/pull/1036#discussion_r739787329
If confirmed, we should bump to at least `opentelemetry-*==1.1.0`, and adjust our OpenTelemetry logic to the changes in the library API. | 1.0 | Bump minimum openetelemetry version to support type checks - See the following comment for details: https://github.com/googleapis/python-bigquery/pull/1036#discussion_r739787329
If confirmed, we should bump to at least `opentelemetry-*==1.1.0`, and adjust our OpenTelemetry logic to the changes in the library API. | process | bump minimum openetelemetry version to support type checks see the following comment for details if confirmed we should bump to at least opentelemetry and adjust our opentelemetry logic to the changes in the library api | 1 |
16,676 | 21,780,412,332 | IssuesEvent | 2022-05-13 18:13:23 | GoogleCloudPlatform/emblem | https://api.github.com/repos/GoogleCloudPlatform/emblem | closed | Proposal: Release Process | type: process persona: maintainer | This issue is created to propose a release process. I want to get some early agreement here, then will convert to a PR documenting the process in more detail.
* Commits will follow [Conventional Commit messages](https://www.conventionalcommits.org/)
* We will try [conventional-commit-lint bot](https://github.com/googleapis/repo-automation-bots/tree/master/packages/conventional-commit-lint) to see if it helps us follow this practice
* Versioning will follow [Semantic Versioning](semver.org).
* Changelog & release automation with [release-please](https://github.com/googleapis/release-please).
* release-please creates a PR to update changelog, tie into this PR for release smoke tests
* Human reviews the release and double-checks we shouldn't wait for any PRs in flight.
* Manual steps to modify the release (automation TBD):
* Add a button to deploy the release's version
* Link to each decision made since the last release
* Link to each User Journey that changed since the last release
* Releases to amplify
* If there are other release materials such as video or blog, cross-link.
* Tweet the blog post | 1.0 | Proposal: Release Process - This issue is created to propose a release process. I want to get some early agreement here, then will convert to a PR documenting the process in more detail.
* Commits will follow [Conventional Commit messages](https://www.conventionalcommits.org/)
* We will try [conventional-commit-lint bot](https://github.com/googleapis/repo-automation-bots/tree/master/packages/conventional-commit-lint) to see if it helps us follow this practice
* Versioning will follow [Semantic Versioning](semver.org).
* Changelog & release automation with [release-please](https://github.com/googleapis/release-please).
* release-please creates a PR to update changelog, tie into this PR for release smoke tests
* Human reviews the release and double-checks we shouldn't wait for any PRs in flight.
* Manual steps to modify the release (automation TBD):
* Add a button to deploy the release's version
* Link to each decision made since the last release
* Link to each User Journey that changed since the last release
* Releases to amplify
* If there are other release materials such as video or blog, cross-link.
* Tweet the blog post | process | proposal release process this issue is created to propose a release process i want to get some early agreement here then will convert to a pr documenting the process in more detail commits will follow we will try to see if it helps us follow this practice versioning will follow semver org changelog release automation with release please creates a pr to update changelog tie into this pr for release smoke tests human reviews the release and double checks we shouldn t wait for any prs in flight manual steps to modify the release automation tbd add a button to deploy the release s version link to each decision made since the last release link to each user journey that changed since the last release releases to amplify if there are other release materials such as video or blog cross link tweet the blog post | 1 |
63,244 | 6,830,489,374 | IssuesEvent | 2017-11-09 07:02:18 | hypermodules/hyperamp | https://api.github.com/repos/hypermodules/hyperamp | closed | Fix calling t.end() twice. | testing | We don't remove event emitter listeners when we end the test or something in the artwork cache which causes inconsistent test failures when t.end gets called twice.
https://github.com/hypermodules/hyperamp/blob/master/main/lib/artwork-cache/test.js | 1.0 | Fix calling t.end() twice. - We don't remove event emitter listeners when we end the test or something in the artwork cache which causes inconsistent test failures when t.end gets called twice.
https://github.com/hypermodules/hyperamp/blob/master/main/lib/artwork-cache/test.js | non_process | fix calling t end twice we don t remove event emitter listeners when we end the test or something in the artwork cache which causes inconsistent test failures when t end gets called twice | 0 |
536,460 | 15,709,369,451 | IssuesEvent | 2021-03-26 22:22:01 | sopra-fs21-group-16/mth-client | https://api.github.com/repos/sopra-fs21-group-16/mth-client | opened | Create a visual clue to inform user of rejection of activity (swiping) | medium priority task | Create a visual clue, such that if an user swipes an suggested activity to the left, the user gets a feedback that indicates that the user has discarded the suggested activity #3. | 1.0 | Create a visual clue to inform user of rejection of activity (swiping) - Create a visual clue, such that if an user swipes an suggested activity to the left, the user gets a feedback that indicates that the user has discarded the suggested activity #3. | non_process | create a visual clue to inform user of rejection of activity swiping create a visual clue such that if an user swipes an suggested activity to the left the user gets a feedback that indicates that the user has discarded the suggested activity | 0 |
52,085 | 6,218,381,489 | IssuesEvent | 2017-07-09 00:54:03 | ProjectSidewalk/SidewalkWebpage | https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage | opened | Can't close label context menus by clicking on the label | Priority: Low Relaunch Testing | You open the context menu by clicking on a label; the mouse-down does nothing, then the mouse-up opens the menu. When trying to close the menu by clicking on the label, the menu closes on mouse down, then opens back up again on mouse-up. Either
1. Users should be able to close the context menu by clicking on the label, in which case mouse-down should do nothing (instead of closing it) and mouse-up should close it (instead of opening it), or
2. Users should *not* be able to close the menu this way, and so both mouse-down and mouse-up should do nothing, instead of closing and opening the menu, respectively.
This is the behavior on the production server as well as the dev server. Since no one has really noticed it before, it must be an edge case and isn't important right now.
Another related issue: just *clicking* on the curb ramp label repeatedly in this way (which opens and closes the menu) increments whatever counter there is for checking if we provide severity ratings. So I click on the label a few times, and the severity rating reminder popup shows up again. | 1.0 | Can't close label context menus by clicking on the label - You open the context menu by clicking on a label; the mouse-down does nothing, then the mouse-up opens the menu. When trying to close the menu by clicking on the label, the menu closes on mouse down, then opens back up again on mouse-up. Either
1. Users should be able to close the context menu by clicking on the label, in which case mouse-down should do nothing (instead of closing it) and mouse-up should close it (instead of opening it), or
2. Users should *not* be able to close the menu this way, and so both mouse-down and mouse-up should do nothing, instead of closing and opening the menu, respectively.
This is the behavior on the production server as well as the dev server. Since no one has really noticed it before, it must be an edge case and isn't important right now.
Another related issue: just *clicking* on the curb ramp label repeatedly in this way (which opens and closes the menu) increments whatever counter there is for checking if we provide severity ratings. So I click on the label a few times, and the severity rating reminder popup shows up again. | non_process | can t close label context menus by clicking on the label you open the context menu by clicking on a label the mouse down does nothing then the mouse up opens the menu when trying to close the menu by clicking on the label the menu closes on mouse down then opens back up again on mouse up either users should be able to close the context menu by clicking on the label in which case mouse down should do nothing instead of closing it and mouse up should close it instead of opening it or users should not be able to close the menu this way and so both mouse down and mouse up should do nothing instead of closing and opening the menu respectively this is the behavior on the production server as well as the dev server since no one has really noticed it before it must be an edge case and isn t important right now another related issue just clicking on the curb ramp label repeatedly in this way which opens and closes the menu increments whatever counter there is for checking if we provide severity ratings so i click on the label a few times and the severity rating reminder popup shows up again | 0 |
15,120 | 18,852,210,389 | IssuesEvent | 2021-11-11 22:38:17 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Failed queries do not show SQL that was executed | .Proposal Querying/Processor | ### Feature requests and proposals
log every sql executed
I cannot see the sql being executed when it takes too long in question query builder. The sql being executed is the key to debug.
- Your browser and the version: Chrome
- Your operating system: OS X
- Your databases: MySQL
- Metabase version: 0.32.0
- Metabase hosting environment: docker
- Metabase internal database: MySQL | 1.0 | Failed queries do not show SQL that was executed - ### Feature requests and proposals
log every sql executed
I cannot see the sql being executed when it takes too long in question query builder. The sql being executed is the key to debug.
- Your browser and the version: Chrome
- Your operating system: OS X
- Your databases: MySQL
- Metabase version: 0.32.0
- Metabase hosting environment: docker
- Metabase internal database: MySQL | process | failed queries do not show sql that was executed feature requests and proposals log every sql executed i cannot see the sql being executed when it takes too long in question query builder the sql being executed is the key to debug your browser and the version chrome your operating system os x your databases mysql metabase version metabase hosting environment docker metabase internal database mysql | 1 |
6,699 | 9,814,731,140 | IssuesEvent | 2019-06-13 10:53:44 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Batch processing error in r.mapcalc.simple | Bug Processing | Author Name: **Daria Svidzinska** (@darsvid)
Original Redmine Issue: [22008](https://issues.qgis.org/issues/22008)
Affected QGIS version: 3.6.2
Redmine category:processing/core
---
I'm trying to perform batch processing for multiple rasters using a conditional expression in r.mapcalc.simple, GRASS 7.6.1.
The expression is:
@if((A==322||A==386||A==834||A==898||A==1346)&&(B<=10000&&B>=-10000),(B*0.0001),null())@
where A is a qa_raster and B is a ndvi_raster.
The approach works fine for a single job, but when I try to process multiple files I constantly get an error.
For your convenience I attach the following files:
* batch_processing_error.txt - log of batch processing with error
* qgis_version_info.txt - information about QGIS version etc
* test_project.zip - a test project with a couple of layers as a data example
---
- [qgis_version_info.txt](https://issues.qgis.org/attachments/download/14893/qgis_version_info.txt) (Daria Svidzinska)
- [batch_processing_error.txt](https://issues.qgis.org/attachments/download/14894/batch_processing_error.txt) (Daria Svidzinska)
- [test_project.zip](https://issues.qgis.org/attachments/download/14895/test_project.zip) (Daria Svidzinska) | 1.0 | Batch processing error in r.mapcalc.simple - Author Name: **Daria Svidzinska** (@darsvid)
Original Redmine Issue: [22008](https://issues.qgis.org/issues/22008)
Affected QGIS version: 3.6.2
Redmine category:processing/core
---
I'm trying to perform batch processing for multiple rasters using a conditional expression in r.mapcalc.simple, GRASS 7.6.1.
The expression is:
@if((A==322||A==386||A==834||A==898||A==1346)&&(B<=10000&&B>=-10000),(B*0.0001),null())@
where A is a qa_raster and B is a ndvi_raster.
The approach works fine for a single job, but when I try to process multiple files I constantly get an error.
For your convenience I attach the following files:
* batch_processing_error.txt - log of batch processing with error
* qgis_version_info.txt - information about QGIS version etc
* test_project.zip - a test project with a couple of layers as a data example
---
- [qgis_version_info.txt](https://issues.qgis.org/attachments/download/14893/qgis_version_info.txt) (Daria Svidzinska)
- [batch_processing_error.txt](https://issues.qgis.org/attachments/download/14894/batch_processing_error.txt) (Daria Svidzinska)
- [test_project.zip](https://issues.qgis.org/attachments/download/14895/test_project.zip) (Daria Svidzinska) | process | batch processing error in r mapcalc simple author name daria svidzinska darsvid original redmine issue affected qgis version redmine category processing core i m trying to perform batch processing for multiple rasters using a conditional expression in r mapcalc simple grass the expression is if a a a a a b b null where a is a qa raster and b is a ndvi raster the approach works fine for a single job but when i try to process multiple files i constantly get an error for your convenience i attach the following files batch processing error txt log of batch processing with error qgis version info txt information about qgis version etc test project zip a test project with a couple of layers as a data example daria svidzinska daria svidzinska daria svidzinska | 1 |
19,010 | 25,010,562,762 | IssuesEvent | 2022-11-03 14:58:37 | hashgraph/hedera-mirror-node | https://api.github.com/repos/hashgraph/hedera-mirror-node | closed | DNS resolution error on Apple M1 | bug process | ### Description
```
Unable to load io.netty.resolver.dns.macos.MacOSDnsServerAddressStreamProvider,fallback to system defaults. This may result in incorrect DNS resolutions on MacOS.
java.lang.ClassNotFoundException: io.netty.resolver.dns.macos.MacOSDnsServerAddressStreamProvider
```
### Steps to reproduce
Run grpc, importer or web3 locally
### Additional context
_No response_
### Hedera network
other
### Version
main
### Operating system
_No response_ | 1.0 | DNS resolution error on Apple M1 - ### Description
```
Unable to load io.netty.resolver.dns.macos.MacOSDnsServerAddressStreamProvider,fallback to system defaults. This may result in incorrect DNS resolutions on MacOS.
java.lang.ClassNotFoundException: io.netty.resolver.dns.macos.MacOSDnsServerAddressStreamProvider
```
### Steps to reproduce
Run grpc, importer or web3 locally
### Additional context
_No response_
### Hedera network
other
### Version
main
### Operating system
_No response_ | process | dns resolution error on apple description unable to load io netty resolver dns macos macosdnsserveraddressstreamprovider fallback to system defaults this may result in incorrect dns resolutions on macos java lang classnotfoundexception io netty resolver dns macos macosdnsserveraddressstreamprovider steps to reproduce run grpc importer or locally additional context no response hedera network other version main operating system no response | 1 |
15,349 | 8,852,554,632 | IssuesEvent | 2019-01-08 18:41:23 | webhintio/hint | https://api.github.com/repos/webhintio/hint | reopened | Add rule for best practices for preload-prefetch-and-priorities | area:hint hint-category:performance | Addy Osmani published an interesting article about Chrome's [preload prefetch and priorities](https://medium.com/reloading/preload-prefetch-and-priorities-in-chrome-776165961bbf)
We should:
* [ ] Investigate how it works in Edge and Firefox.
* [ ] See if we can extrapolate a general best practices based on some patterns. | True | Add rule for best practices for preload-prefetch-and-priorities - Addy Osmani published an interesting article about Chrome's [preload prefetch and priorities](https://medium.com/reloading/preload-prefetch-and-priorities-in-chrome-776165961bbf)
We should:
* [ ] Investigate how it works in Edge and Firefox.
* [ ] See if we can extrapolate a general best practices based on some patterns. | non_process | add rule for best practices for preload prefetch and priorities addy osmani published an interesting article about chrome s we should investigate how it works in edge and firefox see if we can extrapolate a general best practices based on some patterns | 0 |
349,569 | 31,813,135,322 | IssuesEvent | 2023-09-13 18:20:09 | golang/go | https://api.github.com/repos/golang/go | closed | os/exec: TestCommand fails on Windows when `NoDefaultCurrentDirectoryInExePath` is set in the environment | Testing OS-Windows NeedsInvestigation | In investigating the data race reported in https://go.dev/cl/527337, I looked at the existing tests that cover the call to `lookExtensions` is `exec.Cmd.Start` on Windows.
I found several test cases in `TestCommand` that look suspicious: they expect `exec.LookPath` to always resolve a bare filename relative to the current directory, possibly with an `ErrDot` error. However, if `NoDefaultCurrentDirectoryInExePath` is set in the environment, `exec.LookPath` does not even attempt to resolve in the current directory, so the tests presumably won't pass.
Running the tests on a gomote confirms that hypothesis as of https://go.dev/cl/527337:
```
gopher@GOLANG-BUILDLET C:\workdir\go\src>set NoDefaultCurrentDirectoryInExePath=TRUE
gopher@GOLANG-BUILDLET C:\workdir\go\src>..\bin\go test os/exec
--- FAIL: TestCommand (0.58s)
--- FAIL: TestCommand/2 (0.06s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir: arg
0:a want:a.exe fails:false}: want "a.exe", got "p2\\a.exe"
--- FAIL: TestCommand/10 (0.07s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir:p ar
g0:a.exe want:p\a.exe fails:false}: want "p\\a.exe", got "p2\\a.exe"
--- FAIL: TestCommand/11 (0.06s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir:p ar
g0:a want:p\a.exe fails:false}: want "p\\a.exe", got "p2\\a.exe"
--- FAIL: TestCommand/1 (0.06s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir: arg
0:a.exe want:a.exe fails:false}: want "a.exe", got "p2\\a.exe"
FAIL
FAIL os/exec 0.947s
FAIL
```
(CC @golang/windows, @ianlancetaylor) | 1.0 | os/exec: TestCommand fails on Windows when `NoDefaultCurrentDirectoryInExePath` is set in the environment - In investigating the data race reported in https://go.dev/cl/527337, I looked at the existing tests that cover the call to `lookExtensions` is `exec.Cmd.Start` on Windows.
I found several test cases in `TestCommand` that look suspicious: they expect `exec.LookPath` to always resolve a bare filename relative to the current directory, possibly with an `ErrDot` error. However, if `NoDefaultCurrentDirectoryInExePath` is set in the environment, `exec.LookPath` does not even attempt to resolve in the current directory, so the tests presumably won't pass.
Running the tests on a gomote confirms that hypothesis as of https://go.dev/cl/527337:
```
gopher@GOLANG-BUILDLET C:\workdir\go\src>set NoDefaultCurrentDirectoryInExePath=TRUE
gopher@GOLANG-BUILDLET C:\workdir\go\src>..\bin\go test os/exec
--- FAIL: TestCommand (0.58s)
--- FAIL: TestCommand/2 (0.06s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir: arg
0:a want:a.exe fails:false}: want "a.exe", got "p2\\a.exe"
--- FAIL: TestCommand/10 (0.07s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir:p ar
g0:a.exe want:p\a.exe fails:false}: want "p\\a.exe", got "p2\\a.exe"
--- FAIL: TestCommand/11 (0.06s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir:p ar
g0:a want:p\a.exe fails:false}: want "p\\a.exe", got "p2\\a.exe"
--- FAIL: TestCommand/1 (0.06s)
lp_windows_test.go:396: test={PATH:p2;p files:[a.exe p\a.exe p2\a.exe] dir: arg
0:a.exe want:a.exe fails:false}: want "a.exe", got "p2\\a.exe"
FAIL
FAIL os/exec 0.947s
FAIL
```
(CC @golang/windows, @ianlancetaylor) | non_process | os exec testcommand fails on windows when nodefaultcurrentdirectoryinexepath is set in the environment in investigating the data race reported in i looked at the existing tests that cover the call to lookextensions is exec cmd start on windows i found several test cases in testcommand that look suspicious they expect exec lookpath to always resolve a bare filename relative to the current directory possibly with an errdot error however if nodefaultcurrentdirectoryinexepath is set in the environment exec lookpath does not even attempt to resolve in the current directory so the tests presumably won t pass running the tests on a gomote confirms that hypothesis as of gopher golang buildlet c workdir go src set nodefaultcurrentdirectoryinexepath true gopher golang buildlet c workdir go src bin go test os exec fail testcommand fail testcommand lp windows test go test path p files dir arg a want a exe fails false want a exe got a exe fail testcommand lp windows test go test path p files dir p ar a exe want p a exe fails false want p a exe got a exe fail testcommand lp windows test go test path p files dir p ar a want p a exe fails false want p a exe got a exe fail testcommand lp windows test go test path p files dir arg a exe want a exe fails false want a exe got a exe fail fail os exec fail cc golang windows ianlancetaylor | 0 |
16,486 | 21,444,061,565 | IssuesEvent | 2022-04-25 03:02:50 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Multipart to Singlepart feature could not be written | Feedback stale Processing Bug | ### What is the bug or the crash?
After the update to QGIS 3.24.0 Tisler, the 'Multipart to Singleparts' yields errors stating the features could not be written. As a result, no data was written to the result set (memory layer called Single_parts).
In the previous version of QGIS, running Multipart to Singleparts on the same dataset worked like a charm.
```
QGIS version: 3.24.0-Tisler
QGIS code revision: e6ab9473e0
Qt version: 5.15.2
Python version: 3.9.5
GDAL version: 3.3.2
GEOS version: 3.9.1-CAPI-1.14.2
PROJ version: Rel. 8.1.1, September 1st, 2021
PDAL version: 2.3.0 (git-version: Release)
Algorithm started at: 2022-03-02T09:58:56
Algorithm 'Multipart to singleparts' starting…
Input parameters:
{ 'INPUT' : 'postgres://dbname=\'stamdata_nl\' host=127.0.0.1 port=5433 sslmode=disable authcfg=v936t9f key=\'tid\' srid=28992 type=MultiPoint checkPrimaryKeyUnicity=\'1\' table="demandpoints"."pc4" (geopunten_verzameling_primair)', 'OUTPUT' : 'TEMPORARY_OUTPUT' }
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
...
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Message log truncated
```
### Steps to reproduce the issue
1. Go to vector -> Geometrie-gereedschappen -> Multiparts to Singleparts
2. Select a layer of geometry type Point(MultiPoint). In my case:
```
= Information from provider =
Storage | PostgreSQL database with PostGIS extension
Encoding |
Geometry | Point (MultiPoint)
Extent | 13611.5550000000002910,306922.0769999999902211 : 277709.4521417830837891,613137.8767019889783114
Feature count | 4.069
= Coordinate Reference System (CRS) =
Name | EPSG:28992 - Amersfoort / RD New
Units | meters
Method | Oblique Stereographic Alternative
Celestial body | Earth
Reference | Static (relies on a datum which is plate-fixed)
```
3. Click Run.
### Versions
QGIS version | 3.24.0-Tisler | QGIS code revision | e6ab9473e0
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.3.2
PROJ version | 8.1.1
EPSG Registry database version | v10.028 (2021-07-07)
GEOS version | 3.9.1-CAPI-1.14.2
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.6
QScintilla2 version | 2.11.5
OS version | macOS 12.2
| | |
Active Python plugins
contour | 2.0.8
processing | 2.12.99
sagaprovider | 2.12.99
grassprovider | 2.12.99
db_manager | 0.1.20
MetaSearch | 0.3.6
QGIS version
3.24.0-Tisler
QGIS code revision
[e6ab9473e0](https://github.com/qgis/QGIS/commit/e6ab9473e0)
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.3.2
PROJ version
8.1.1
EPSG Registry database version
v10.028 (2021-07-07)
GEOS version
3.9.1-CAPI-1.14.2
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.6
QScintilla2 version
2.11.5
OS version
macOS 12.2
Active Python plugins
contour
2.0.8
processing
2.12.99
sagaprovider
2.12.99
grassprovider
2.12.99
db_manager
0.1.20
MetaSearch
0.3.6
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_ | 1.0 | Multipart to Singlepart feature could not be written - ### What is the bug or the crash?
After the update to QGIS 3.24.0 Tisler, the 'Multipart to Singleparts' yields errors stating the features could not be written. As a result, no data was written to the result set (memory layer called Single_parts).
In the previous version of QGIS, running Multipart to Singleparts on the same dataset worked like a charm.
```
QGIS version: 3.24.0-Tisler
QGIS code revision: e6ab9473e0
Qt version: 5.15.2
Python version: 3.9.5
GDAL version: 3.3.2
GEOS version: 3.9.1-CAPI-1.14.2
PROJ version: Rel. 8.1.1, September 1st, 2021
PDAL version: 2.3.0 (git-version: Release)
Algorithm started at: 2022-03-02T09:58:56
Algorithm 'Multipart to singleparts' starting…
Input parameters:
{ 'INPUT' : 'postgres://dbname=\'stamdata_nl\' host=127.0.0.1 port=5433 sslmode=disable authcfg=v936t9f key=\'tid\' srid=28992 type=MultiPoint checkPrimaryKeyUnicity=\'1\' table="demandpoints"."pc4" (geopunten_verzameling_primair)', 'OUTPUT' : 'TEMPORARY_OUTPUT' }
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
...
Feature could not be written to Single_parts_45aa0dae_29f2_445d_a6ef_55b4c2f1baf5: Could not store attribute "geopunten_verzameling_volledig": Could not convert value "" to target type
Message log truncated
```
### Steps to reproduce the issue
1. Go to vector -> Geometrie-gereedschappen -> Multiparts to Singleparts
2. Select a layer of geometry type Point(MultiPoint). In my case:
```
= Information from provider =
Storage | PostgreSQL database with PostGIS extension
Encoding |
Geometry | Point (MultiPoint)
Extent | 13611.5550000000002910,306922.0769999999902211 : 277709.4521417830837891,613137.8767019889783114
Feature count | 4.069
= Coordinate Reference System (CRS) =
Name | EPSG:28992 - Amersfoort / RD New
Units | meters
Method | Oblique Stereographic Alternative
Celestial body | Earth
Reference | Static (relies on a datum which is plate-fixed)
```
3. Click Run.
### Versions
QGIS version | 3.24.0-Tisler | QGIS code revision | e6ab9473e0
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.3.2
PROJ version | 8.1.1
EPSG Registry database version | v10.028 (2021-07-07)
GEOS version | 3.9.1-CAPI-1.14.2
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.6
QScintilla2 version | 2.11.5
OS version | macOS 12.2
| | |
Active Python plugins
contour | 2.0.8
processing | 2.12.99
sagaprovider | 2.12.99
grassprovider | 2.12.99
db_manager | 0.1.20
MetaSearch | 0.3.6
QGIS version
3.24.0-Tisler
QGIS code revision
[e6ab9473e0](https://github.com/qgis/QGIS/commit/e6ab9473e0)
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.3.2
PROJ version
8.1.1
EPSG Registry database version
v10.028 (2021-07-07)
GEOS version
3.9.1-CAPI-1.14.2
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.6
QScintilla2 version
2.11.5
OS version
macOS 12.2
Active Python plugins
contour
2.0.8
processing
2.12.99
sagaprovider
2.12.99
grassprovider
2.12.99
db_manager
0.1.20
MetaSearch
0.3.6
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_ | process | multipart to singlepart feature could not be written what is the bug or the crash after the update to qgis tisler the multipart to singleparts yields errors stating the features could not be written as a result no data was written to the result set memory layer called single parts in the previous version of qgis running multipart to singleparts on the same dataset worked like a charm qgis version tisler qgis code revision qt version python version gdal version geos version capi proj version rel september pdal version git version release algorithm started at algorithm multipart to singleparts starting… input parameters input postgres dbname stamdata nl host port sslmode disable authcfg key tid srid type multipoint checkprimarykeyunicity table demandpoints geopunten verzameling primair output temporary output feature could not be written to single parts could not store attribute geopunten verzameling volledig could not convert value to target type feature could not be written to single parts could not store attribute geopunten verzameling volledig could not convert value to target type feature could not be written to single parts could not store attribute geopunten verzameling volledig could not convert value to target type feature could not be written to single parts could not store attribute geopunten verzameling volledig could not convert value to target type feature could not be written to single parts could not store attribute geopunten verzameling volledig could not convert value to target type message log truncated steps to reproduce the issue go to vector geometrie gereedschappen multiparts to singleparts select a layer of geometry type point multipoint in my case information from provider storage postgresql database with postgis extension encoding geometry point multipoint extent feature count coordinate reference system crs name epsg amersfoort rd new units meters method oblique stereographic alternative celestial body earth reference static relies on a datum which is plate fixed click run versions qgis version tisler qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version macos active python plugins contour processing sagaprovider grassprovider db manager metasearch qgis version tisler qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version macos active python plugins contour processing sagaprovider grassprovider db manager metasearch supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response | 1 |
5,020 | 7,845,548,341 | IssuesEvent | 2018-06-19 13:16:02 | AffiliateWP/AffiliateWP | https://api.github.com/repos/AffiliateWP/AffiliateWP | opened | This error is showing up on the referrals table | add-on batch-processing bug has HS to-notify | A customer is experiencing the following error:
>Fatal error: Uncaught Error: Call to a member function type() on boolean in /home/s793/html/wp-content/plugins/AffiliateWP-2.2.3/includes/admin/referrals/class-list-table.php:251 Stack trace: #0 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1254): AffWP_Referrals_Table->column_default(false, 'type') #1 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1192): WP_List_Table->single_row_columns(false) #2 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1180): WP_List_Table->single_row(false) #3 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1165): WP_List_Table->display_rows() #4 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1097): WP_List_Table->display_rows_or_placeholder() #5 /home/s793/html/wp-content/plugins/AffiliateWP-2.2.3/includes/admin/referrals/referrals.php(99): WP_List_Table->display() #6 /home/s793/html/wp-includes/class-wp-hook.php(286): affwp_referrals_admin('') #7 /home/s793/html/wp-includes/class-wp-hook.php(310): WP_Hook->apply_filters('', Array) #8 /home in /home/s793/html/wp-content/plugins/AffiliateWP-2.2.3/includes/admin/referrals/class-list-table.php on line 251
I have not been able to replicate this issue yet.
Ticket: https://secure.helpscout.net/conversation/600924552/84152?folderId=634609. | 1.0 | This error is showing up on the referrals table - A customer is experiencing the following error:
>Fatal error: Uncaught Error: Call to a member function type() on boolean in /home/s793/html/wp-content/plugins/AffiliateWP-2.2.3/includes/admin/referrals/class-list-table.php:251 Stack trace: #0 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1254): AffWP_Referrals_Table->column_default(false, 'type') #1 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1192): WP_List_Table->single_row_columns(false) #2 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1180): WP_List_Table->single_row(false) #3 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1165): WP_List_Table->display_rows() #4 /home/s793/html/wp-admin/includes/class-wp-list-table.php(1097): WP_List_Table->display_rows_or_placeholder() #5 /home/s793/html/wp-content/plugins/AffiliateWP-2.2.3/includes/admin/referrals/referrals.php(99): WP_List_Table->display() #6 /home/s793/html/wp-includes/class-wp-hook.php(286): affwp_referrals_admin('') #7 /home/s793/html/wp-includes/class-wp-hook.php(310): WP_Hook->apply_filters('', Array) #8 /home in /home/s793/html/wp-content/plugins/AffiliateWP-2.2.3/includes/admin/referrals/class-list-table.php on line 251
I have not been able to replicate this issue yet.
Ticket: https://secure.helpscout.net/conversation/600924552/84152?folderId=634609. | process | this error is showing up on the referrals table a customer is experiencing the following error fatal error uncaught error call to a member function type on boolean in home html wp content plugins affiliatewp includes admin referrals class list table php stack trace home html wp admin includes class wp list table php affwp referrals table column default false type home html wp admin includes class wp list table php wp list table single row columns false home html wp admin includes class wp list table php wp list table single row false home html wp admin includes class wp list table php wp list table display rows home html wp admin includes class wp list table php wp list table display rows or placeholder home html wp content plugins affiliatewp includes admin referrals referrals php wp list table display home html wp includes class wp hook php affwp referrals admin home html wp includes class wp hook php wp hook apply filters array home in home html wp content plugins affiliatewp includes admin referrals class list table php on line i have not been able to replicate this issue yet ticket | 1 |
16,206 | 20,732,470,540 | IssuesEvent | 2022-03-14 10:41:33 | prisma/prisma | https://api.github.com/repos/prisma/prisma | closed | `'then' in PrimsaPromise` returns false | bug/2-confirmed kind/bug process/candidate topic: prisma-client tech/typescript team/client topic: fluent api | ### Bug description
Some libraries use `'then' in x` as part of their detection for promise-like/thenable objects.
The way prisma client builds its fluent APIs with proxies causes this to fail.
### How to reproduce
`console.log('then' in prisma.someModel.findUnique({ where: { id: 1 }))`
`console.log(typeof prisma.someModel.findUnique({ where: { id: 1 })).then === 'function')`
This is basically because ownKeys here https://github.com/prisma/prisma/blob/3855438b964b40edeb72a3815c641d97f8c433c4/packages/client/src/runtime/core/model/applyFluent.ts#L106 does not include `then` (or other promise methods)
### Expected behavior
`'then' in promise` should return true.
### Prisma information
<!-- Do not include your database credentials when sharing your Prisma schema! -->
### Environment & setup
- OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]-->
- Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]-->
- Node.js version: <!--[Run `node -v` to see your Node.js version]-->
### Prisma Version
```
prisma : 3.10.0
@prisma/client : 3.10.0
Current platform : darwin
Query Engine (Node-API) : libquery-engine 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 73e60b76d394f8d37d8ebd1f8918c79029f0db86
Studio : 0.458.0
```
| 1.0 | `'then' in PrimsaPromise` returns false - ### Bug description
Some libraries use `'then' in x` as part of their detection for promise-like/thenable objects.
The way prisma client builds its fluent APIs with proxies causes this to fail.
### How to reproduce
`console.log('then' in prisma.someModel.findUnique({ where: { id: 1 }))`
`console.log(typeof prisma.someModel.findUnique({ where: { id: 1 })).then === 'function')`
This is basically because ownKeys here https://github.com/prisma/prisma/blob/3855438b964b40edeb72a3815c641d97f8c433c4/packages/client/src/runtime/core/model/applyFluent.ts#L106 does not include `then` (or other promise methods)
### Expected behavior
`'then' in promise` should return true.
### Prisma information
<!-- Do not include your database credentials when sharing your Prisma schema! -->
### Environment & setup
- OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]-->
- Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]-->
- Node.js version: <!--[Run `node -v` to see your Node.js version]-->
### Prisma Version
```
prisma : 3.10.0
@prisma/client : 3.10.0
Current platform : darwin
Query Engine (Node-API) : libquery-engine 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 73e60b76d394f8d37d8ebd1f8918c79029f0db86 (at ../../node_modules/.pnpm/@prisma+engines@3.10.0-50.73e60b76d394f8d37d8ebd1f8918c79029f0db86/node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 73e60b76d394f8d37d8ebd1f8918c79029f0db86
Studio : 0.458.0
```
| process | then in primsapromise returns false bug description some libraries use then in x as part of their detection for promise like thenable objects the way prisma client builds its fluent apis with proxies causes this to fail how to reproduce console log then in prisma somemodel findunique where id console log typeof prisma somemodel findunique where id then function this is basically because ownkeys here does not include then or other promise methods expected behavior then in promise should return true prisma information environment setup os database node js version prisma version prisma prisma client current platform darwin query engine node api libquery engine at node modules pnpm prisma engines node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at node modules pnpm prisma engines node modules prisma engines migration engine darwin introspection engine introspection core at node modules pnpm prisma engines node modules prisma engines introspection engine darwin format binary prisma fmt at node modules pnpm prisma engines node modules prisma engines prisma fmt darwin default engines hash studio | 1 |
2,962 | 5,959,741,134 | IssuesEvent | 2017-05-29 12:05:05 | itsyouonline/identityserver | https://api.github.com/repos/itsyouonline/identityserver | closed | GET request does not work for acquiring a JWT | process_wontfix type_bug | According to the docs, to acquire a JWT, one should do a GET request like:
`curl -H "Authorization: token OAUTH-TOKEN" https://itsyou.online/v1/oauth/jwt?scope=user:memberof:org1`
This, however, didn't work. Only the POST request did.
```
data = {'aud': 'aliorg', 'scope': 'user:email:main,user:memberof:aliorg'}
headers = {'Authorization': 'token %s' % ouath_token}
url = 'https://itsyou.online/v1/oauth/jwt'
r = requests.get(url, params=data, headers=headers)
print(r.status_code)
Out[59]: 401
```
WIth POST:
```
r = requests.post(url, data=json.dumps(data), headers=headers)
print(r.status_code)
Out[60]: 200
```
```
| 1.0 | GET request does not work for acquiring a JWT - According to the docs, to acquire a JWT, one should do a GET request like:
`curl -H "Authorization: token OAUTH-TOKEN" https://itsyou.online/v1/oauth/jwt?scope=user:memberof:org1`
This, however, didn't work. Only the POST request did.
```
data = {'aud': 'aliorg', 'scope': 'user:email:main,user:memberof:aliorg'}
headers = {'Authorization': 'token %s' % ouath_token}
url = 'https://itsyou.online/v1/oauth/jwt'
r = requests.get(url, params=data, headers=headers)
print(r.status_code)
Out[59]: 401
```
WIth POST:
```
r = requests.post(url, data=json.dumps(data), headers=headers)
print(r.status_code)
Out[60]: 200
```
```
| process | get request does not work for acquiring a jwt according to the docs to acquire a jwt one should do a get request like curl h authorization token oauth token this however didn t work only the post request did data aud aliorg scope user email main user memberof aliorg headers authorization token s ouath token url r requests get url params data headers headers print r status code out with post r requests post url data json dumps data headers headers print r status code out | 1 |
53,054 | 13,260,851,432 | IssuesEvent | 2020-08-20 18:52:11 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | icetray/trunk/resources/docs/i3frame.rst clean up (Trac #636) | IceTray Migrated from Trac defect | needs some serious help
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/636">https://code.icecube.wisc.edu/projects/icecube/ticket/636</a>, reported by anonymousand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:57",
"_ts": "1416713877111216",
"description": "needs some serious help",
"reporter": "anonymous",
"cc": "",
"resolution": "fixed",
"time": "2011-05-19T02:00:44",
"component": "IceTray",
"summary": "icetray/trunk/resources/docs/i3frame.rst clean up",
"priority": "normal",
"keywords": "documentation",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | icetray/trunk/resources/docs/i3frame.rst clean up (Trac #636) - needs some serious help
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/636">https://code.icecube.wisc.edu/projects/icecube/ticket/636</a>, reported by anonymousand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:57",
"_ts": "1416713877111216",
"description": "needs some serious help",
"reporter": "anonymous",
"cc": "",
"resolution": "fixed",
"time": "2011-05-19T02:00:44",
"component": "IceTray",
"summary": "icetray/trunk/resources/docs/i3frame.rst clean up",
"priority": "normal",
"keywords": "documentation",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| non_process | icetray trunk resources docs rst clean up trac needs some serious help migrated from json status closed changetime ts description needs some serious help reporter anonymous cc resolution fixed time component icetray summary icetray trunk resources docs rst clean up priority normal keywords documentation milestone owner troy type defect | 0 |
227,103 | 7,526,993,566 | IssuesEvent | 2018-04-13 15:35:57 | googleapis/nodejs-pubsub | https://api.github.com/repos/googleapis/nodejs-pubsub | closed | Duplicate messages exactly every 15 minutes | priority: p2 status: blocked type: bug | #### Environment details
- OS: Kubernetes Engine
- Node.js version: 6.3.0
- npm version: –
- @google-cloud/pubsub version: 0.16.2
#### Steps to reproduce
This is most likely related to the discussion in https://github.com/googleapis/nodejs-pubsub/issues/2#issuecomment-356423284 (and following comments in that thread). However since the discussion there somewhat faded out I want to report my findings in a new issue.
My subscriber is consuming messages at a rate of roughly 500/s and it is receiving small batches of duplicate messages exactly every 15 minutes. Those batches typically contain between 100 and 400 duplicate messages. Here's a plot of the number of duplicates over time:
<img width="760" alt="batches" src="https://user-images.githubusercontent.com/508118/36157179-8a42208a-10d9-11e8-97d0-2bb00999e2ec.png">
Most of the duplicates are being delivered to my subscriber within less then a second. Here's a histogram of the durations between redeliveries in milliseconds:
<img width="805" alt="redelivery" src="https://user-images.githubusercontent.com/508118/36156975-f7eb5792-10d8-11e8-854b-6d87c60a6916.png">
As you can see, the batches of duplicates coincide with spikes in Stackdriver graphs on `StreamingPull Operations` and `StreamingPull Acknowledge Requests` (please note that Stackdriver shows Berlin time while the above graph shows UTC, hence 1h time difference):
<img width="1095" alt="stackdriver" src="https://user-images.githubusercontent.com/508118/36157278-b9d0f466-10d9-11e8-98fb-a626ef4a196e.png">
[From the comments in the other thread](https://github.com/googleapis/nodejs-pubsub/issues/2#issuecomment-356423284) I did not really understand whether the behavior we see is actually expected. What's the reason for this to happen precisely every 15 minutes?
Even though the absolute number of duplicates is well below 1%, this still looks pretty odd, unexpected and unnecessary. I'd love to understand better what's causing this issue and how it could potentially be fixed.
/cc @kir-titievsky @callmehiphop @rossj | 1.0 | Duplicate messages exactly every 15 minutes - #### Environment details
- OS: Kubernetes Engine
- Node.js version: 6.3.0
- npm version: –
- @google-cloud/pubsub version: 0.16.2
#### Steps to reproduce
This is most likely related to the discussion in https://github.com/googleapis/nodejs-pubsub/issues/2#issuecomment-356423284 (and following comments in that thread). However since the discussion there somewhat faded out I want to report my findings in a new issue.
My subscriber is consuming messages at a rate of roughly 500/s and it is receiving small batches of duplicate messages exactly every 15 minutes. Those batches typically contain between 100 and 400 duplicate messages. Here's a plot of the number of duplicates over time:
<img width="760" alt="batches" src="https://user-images.githubusercontent.com/508118/36157179-8a42208a-10d9-11e8-97d0-2bb00999e2ec.png">
Most of the duplicates are being delivered to my subscriber within less then a second. Here's a histogram of the durations between redeliveries in milliseconds:
<img width="805" alt="redelivery" src="https://user-images.githubusercontent.com/508118/36156975-f7eb5792-10d8-11e8-854b-6d87c60a6916.png">
As you can see, the batches of duplicates coincide with spikes in Stackdriver graphs on `StreamingPull Operations` and `StreamingPull Acknowledge Requests` (please note that Stackdriver shows Berlin time while the above graph shows UTC, hence 1h time difference):
<img width="1095" alt="stackdriver" src="https://user-images.githubusercontent.com/508118/36157278-b9d0f466-10d9-11e8-98fb-a626ef4a196e.png">
[From the comments in the other thread](https://github.com/googleapis/nodejs-pubsub/issues/2#issuecomment-356423284) I did not really understand whether the behavior we see is actually expected. What's the reason for this to happen precisely every 15 minutes?
Even though the absolute number of duplicates is well below 1%, this still looks pretty odd, unexpected and unnecessary. I'd love to understand better what's causing this issue and how it could potentially be fixed.
/cc @kir-titievsky @callmehiphop @rossj | non_process | duplicate messages exactly every minutes environment details os kubernetes engine node js version npm version – google cloud pubsub version steps to reproduce this is most likely related to the discussion in and following comments in that thread however since the discussion there somewhat faded out i want to report my findings in a new issue my subscriber is consuming messages at a rate of roughly s and it is receiving small batches of duplicate messages exactly every minutes those batches typically contain between and duplicate messages here s a plot of the number of duplicates over time img width alt batches src most of the duplicates are being delivered to my subscriber within less then a second here s a histogram of the durations between redeliveries in milliseconds img width alt redelivery src as you can see the batches of duplicates coincide with spikes in stackdriver graphs on streamingpull operations and streamingpull acknowledge requests please note that stackdriver shows berlin time while the above graph shows utc hence time difference img width alt stackdriver src i did not really understand whether the behavior we see is actually expected what s the reason for this to happen precisely every minutes even though the absolute number of duplicates is well below this still looks pretty odd unexpected and unnecessary i d love to understand better what s causing this issue and how it could potentially be fixed cc kir titievsky callmehiphop rossj | 0 |
4,713 | 7,551,185,302 | IssuesEvent | 2018-04-18 19:17:19 | nodejs/node | https://api.github.com/repos/nodejs/node | opened | test-child-process-exec-kill-throws fails intermittently on multiple platforms | child_process | <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: 10.0.0-pre (master)
* **Platform**: n/a
* **Subsystem**: child_process test
<!-- Enter your issue details below this comment. -->
This is a bit disconcerting. Both are from the last hour or two. It seems to have popped up only in the last day or two, so it may be caused by something that landed recently?
Here it is timing out on AIX:
https://ci.nodejs.org/job/node-test-commit-aix/14317/nodes=aix61-ppc64/console
```console
not ok 357 parallel/test-child-process-exec-kill-throws
---
duration_ms: 120.232
severity: fail
exitcode: -15
stack: |-
timeout
```
Here it is timing out on macOS:
https://ci.nodejs.org/job/node-test-commit-osx/17926/nodes=osx1010/console
```console
not ok 548 parallel/test-child-process-exec-kill-throws
---
duration_ms: 120.110
severity: fail
exitcode: -15
stack: |-
timeout
```
@nodejs/child_process | 1.0 | test-child-process-exec-kill-throws fails intermittently on multiple platforms - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: 10.0.0-pre (master)
* **Platform**: n/a
* **Subsystem**: child_process test
<!-- Enter your issue details below this comment. -->
This is a bit disconcerting. Both are from the last hour or two. It seems to have popped up only in the last day or two, so it may be caused by something that landed recently?
Here it is timing out on AIX:
https://ci.nodejs.org/job/node-test-commit-aix/14317/nodes=aix61-ppc64/console
```console
not ok 357 parallel/test-child-process-exec-kill-throws
---
duration_ms: 120.232
severity: fail
exitcode: -15
stack: |-
timeout
```
Here it is timing out on macOS:
https://ci.nodejs.org/job/node-test-commit-osx/17926/nodes=osx1010/console
```console
not ok 548 parallel/test-child-process-exec-kill-throws
---
duration_ms: 120.110
severity: fail
exitcode: -15
stack: |-
timeout
```
@nodejs/child_process | process | test child process exec kill throws fails intermittently on multiple platforms thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version pre master platform n a subsystem child process test this is a bit disconcerting both are from the last hour or two it seems to have popped up only in the last day or two so it may be caused by something that landed recently here it is timing out on aix console not ok parallel test child process exec kill throws duration ms severity fail exitcode stack timeout here it is timing out on macos console not ok parallel test child process exec kill throws duration ms severity fail exitcode stack timeout nodejs child process | 1 |
16,410 | 21,191,417,539 | IssuesEvent | 2022-04-08 17:52:51 | cypress-io/cypress | https://api.github.com/repos/cypress-io/cypress | closed | fix flaky session_spec in server-e2e-tests-electron | process: flaky test stage: icebox | ### Current behavior
example failure: https://app.circleci.com/pipelines/github/cypress-io/cypress/22975/workflows/4a639267-6eda-4e78-84b6-5c1f3426f04e/jobs/845803/parallel-runs/7
### Desired behavior
n/a
### Test code to reproduce
n/a
### Cypress Version
*
### Other
_No response_ | 1.0 | fix flaky session_spec in server-e2e-tests-electron - ### Current behavior
example failure: https://app.circleci.com/pipelines/github/cypress-io/cypress/22975/workflows/4a639267-6eda-4e78-84b6-5c1f3426f04e/jobs/845803/parallel-runs/7
### Desired behavior
n/a
### Test code to reproduce
n/a
### Cypress Version
*
### Other
_No response_ | process | fix flaky session spec in server tests electron current behavior example failure desired behavior n a test code to reproduce n a cypress version other no response | 1 |
332,124 | 29,185,751,683 | IssuesEvent | 2023-05-19 15:17:23 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix jax_lax_operators.test_jax_lax_reshape | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix jax_lax_operators.test_jax_lax_reshape - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4570376767/jobs/8067642702" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
| non_process | fix jax lax operators test jax lax reshape tensorflow img src torch img src numpy img src jax img src | 0 |
19,646 | 26,006,136,126 | IssuesEvent | 2022-12-20 19:34:07 | GSA/EDX | https://api.github.com/repos/GSA/EDX | closed | Onboarding GSA Website Manager - vltp.gsa.gov | process Proposed Next Projects | Welcome to the GSA.
As a Website Manager, responsibilities include:
* A number of conditions, described 21st Century IDEA
### Resources and trainings
* Digital.gov links
### How to get help
Contact EDX
### Information about the Website Inventory
https://touchpoints.app.cloud.gov/admin/websites/591
| 1.0 | Onboarding GSA Website Manager - vltp.gsa.gov - Welcome to the GSA.
As a Website Manager, responsibilities include:
* A number of conditions, described 21st Century IDEA
### Resources and trainings
* Digital.gov links
### How to get help
Contact EDX
### Information about the Website Inventory
https://touchpoints.app.cloud.gov/admin/websites/591
| process | onboarding gsa website manager vltp gsa gov welcome to the gsa as a website manager responsibilities include a number of conditions described century idea resources and trainings digital gov links how to get help contact edx information about the website inventory | 1 |
13,261 | 15,729,686,275 | IssuesEvent | 2021-03-29 15:06:42 | gfx-rs/naga | https://api.github.com/repos/gfx-rs/naga | closed | Validation function for the IR | area: processing help wanted kind: feature | The `Module` struct that our front-ends produce generally has some level of type-level encoding of the constraints, however many properties can only be checked at run-time. One big area of validation is type checking:
- builtins have the expected types
- constructors are using consistent values, e.g. #43
- types generally agree between producer and consumer
Another area is access checking:
- all the handles are actually valid
- input globals can only be read
- can't linearly sample integer textures
- depth images can only be used with comparison samplers
- and more!
Finally we need to enforce some of the conventions. For example, variables can either be stored as pointers (that are loaded/stored), or directly. There is currently no consistency, and whatever we choose would need to be enforced.
We have a [proc](https://github.com/gfx-rs/naga/tree/63ab925f9c0e5e39466863ce70ac002fe4a5c7c8/src/proc) module for end-point-independent processing. I think the `validator` sub-module could live there. The main task of it would be getting a module and making sure that it's valid. We'd be running it in CI tests to make sure that the module contents produced by the front-ends, or by different processors/transformers, are all valid. | 1.0 | Validation function for the IR - The `Module` struct that our front-ends produce generally has some level of type-level encoding of the constraints, however many properties can only be checked at run-time. One big area of validation is type checking:
- builtins have the expected types
- constructors are using consistent values, e.g. #43
- types generally agree between producer and consumer
Another area is access checking:
- all the handles are actually valid
- input globals can only be read
- can't linearly sample integer textures
- depth images can only be used with comparison samplers
- and more!
Finally we need to enforce some of the conventions. For example, variables can either be stored as pointers (that are loaded/stored), or directly. There is currently no consistency, and whatever we choose would need to be enforced.
We have a [proc](https://github.com/gfx-rs/naga/tree/63ab925f9c0e5e39466863ce70ac002fe4a5c7c8/src/proc) module for end-point-independent processing. I think the `validator` sub-module could live there. The main task of it would be getting a module and making sure that it's valid. We'd be running it in CI tests to make sure that the module contents produced by the front-ends, or by different processors/transformers, are all valid. | process | validation function for the ir the module struct that our front ends produce generally has some level of type level encoding of the constraints however many properties can only be checked at run time one big area of validation is type checking builtins have the expected types constructors are using consistent values e g types generally agree between producer and consumer another area is access checking all the handles are actually valid input globals can only be read can t linearly sample integer textures depth images can only be used with comparison samplers and more finally we need to enforce some of the conventions for example variables can either be stored as pointers that are loaded stored or directly there is currently no consistency and whatever we choose would need to be enforced we have a module for end point independent processing i think the validator sub module could live there the main task of it would be getting a module and making sure that it s valid we d be running it in ci tests to make sure that the module contents produced by the front ends or by different processors transformers are all valid | 1 |
6,732 | 9,854,683,236 | IssuesEvent | 2019-06-19 17:27:56 | googleapis/google-cloud-python | https://api.github.com/repos/googleapis/google-cloud-python | reopened | Bigtable: 'test_create_instance_w_two_clusters' flakes with '504 Deadline Exceeded' | api: bigtable flaky testing type: process | /cc @sduskis, @vikas-jamdar
From:
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7977
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7978
```python
___________ TestInstanceAdminAPI.test_create_instance_w_two_clusters ___________
self = <tests.system.TestInstanceAdminAPI testMethod=test_create_instance_w_two_clusters>
def test_create_instance_w_two_clusters(self):
from google.cloud.bigtable import enums
from google.cloud.bigtable.table import ClusterState
_PRODUCTION = enums.Instance.Type.PRODUCTION
ALT_INSTANCE_ID = 'dif' + unique_resource_id('-')
instance = Config.CLIENT.instance(ALT_INSTANCE_ID,
instance_type=_PRODUCTION,
labels=LABELS)
ALT_CLUSTER_ID_1 = ALT_INSTANCE_ID + '-c1'
ALT_CLUSTER_ID_2 = ALT_INSTANCE_ID + '-c2'
LOCATION_ID_2 = 'us-central1-f'
STORAGE_TYPE = enums.StorageType.HDD
cluster_1 = instance.cluster(
ALT_CLUSTER_ID_1, location_id=LOCATION_ID, serve_nodes=SERVE_NODES,
default_storage_type=STORAGE_TYPE)
cluster_2 = instance.cluster(
ALT_CLUSTER_ID_2, location_id=LOCATION_ID_2,
serve_nodes=SERVE_NODES, default_storage_type=STORAGE_TYPE)
operation = instance.create(clusters=[cluster_1, cluster_2])
# We want to make sure the operation completes.
operation.result(timeout=10)
# Make sure this instance gets deleted after the test case.
self.instances_to_delete.append(instance)
# Create a new instance instance and make sure it is the same.
instance_alt = Config.CLIENT.instance(ALT_INSTANCE_ID)
instance_alt.reload()
self.assertEqual(instance, instance_alt)
self.assertEqual(instance.display_name, instance_alt.display_name)
self.assertEqual(instance.type_, instance_alt.type_)
clusters, failed_locations = instance_alt.list_clusters()
self.assertEqual(failed_locations, [])
clusters.sort(key=lambda x: x.name)
alt_cluster_1, alt_cluster_2 = clusters
self.assertEqual(cluster_1.location_id, alt_cluster_1.location_id)
self.assertEqual(alt_cluster_1.state, enums.Cluster.State.READY)
self.assertEqual(cluster_1.serve_nodes, alt_cluster_1.serve_nodes)
self.assertEqual(cluster_1.default_storage_type,
alt_cluster_1.default_storage_type)
self.assertEqual(cluster_2.location_id, alt_cluster_2.location_id)
self.assertEqual(alt_cluster_2.state, enums.Cluster.State.READY)
self.assertEqual(cluster_2.serve_nodes, alt_cluster_2.serve_nodes)
self.assertEqual(cluster_2.default_storage_type,
alt_cluster_2.default_storage_type)
# Test list clusters in project via 'client.list_clusters'
clusters, failed_locations = Config.CLIENT.list_clusters()
self.assertFalse(failed_locations)
found = set([cluster.name for cluster in clusters])
self.assertTrue({alt_cluster_1.name,
alt_cluster_2.name,
Config.CLUSTER.name}.issubset(found))
temp_table_id = 'test-get-cluster-states'
temp_table = instance.table(temp_table_id)
> temp_table.create()
tests/system.py:280:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/bigtable/table.py:218: in create
table=table, initial_splits=splits)
google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py:327: in create_table
request, retry=retry, timeout=timeout, metadata=metadata)
../api_core/google/api_core/gapic_v1/method.py:139: in __call__
return wrapped_func(*args, **kwargs)
../api_core/google/api_core/retry.py:260: in retry_wrapped_func
on_error=on_error,
../api_core/google/api_core/retry.py:177: in retry_target
return target()
../api_core/google/api_core/timeout.py:206: in func_with_timeout
return func(*args, **kwargs)
../api_core/google/api_core/grpc_helpers.py:61: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = DeadlineExceeded('Deadline Exceeded',)
from_value = <_Rendezvous of RPC that terminated with:
status = StatusCode.DEADLINE_EXCEED...all.cc","file_line":1099,"grpc_message":"Deadline Exceeded","grpc_status":4}"
>
def raise_from(value, from_value):
> raise value
E DeadlineExceeded: 504 Deadline Exceeded
../.nox/sys-2-7/lib/python2.7/site-packages/six.py:737: DeadlineExceeded
``` | 1.0 | Bigtable: 'test_create_instance_w_two_clusters' flakes with '504 Deadline Exceeded' - /cc @sduskis, @vikas-jamdar
From:
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7977
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7978
```python
___________ TestInstanceAdminAPI.test_create_instance_w_two_clusters ___________
self = <tests.system.TestInstanceAdminAPI testMethod=test_create_instance_w_two_clusters>
def test_create_instance_w_two_clusters(self):
from google.cloud.bigtable import enums
from google.cloud.bigtable.table import ClusterState
_PRODUCTION = enums.Instance.Type.PRODUCTION
ALT_INSTANCE_ID = 'dif' + unique_resource_id('-')
instance = Config.CLIENT.instance(ALT_INSTANCE_ID,
instance_type=_PRODUCTION,
labels=LABELS)
ALT_CLUSTER_ID_1 = ALT_INSTANCE_ID + '-c1'
ALT_CLUSTER_ID_2 = ALT_INSTANCE_ID + '-c2'
LOCATION_ID_2 = 'us-central1-f'
STORAGE_TYPE = enums.StorageType.HDD
cluster_1 = instance.cluster(
ALT_CLUSTER_ID_1, location_id=LOCATION_ID, serve_nodes=SERVE_NODES,
default_storage_type=STORAGE_TYPE)
cluster_2 = instance.cluster(
ALT_CLUSTER_ID_2, location_id=LOCATION_ID_2,
serve_nodes=SERVE_NODES, default_storage_type=STORAGE_TYPE)
operation = instance.create(clusters=[cluster_1, cluster_2])
# We want to make sure the operation completes.
operation.result(timeout=10)
# Make sure this instance gets deleted after the test case.
self.instances_to_delete.append(instance)
# Create a new instance instance and make sure it is the same.
instance_alt = Config.CLIENT.instance(ALT_INSTANCE_ID)
instance_alt.reload()
self.assertEqual(instance, instance_alt)
self.assertEqual(instance.display_name, instance_alt.display_name)
self.assertEqual(instance.type_, instance_alt.type_)
clusters, failed_locations = instance_alt.list_clusters()
self.assertEqual(failed_locations, [])
clusters.sort(key=lambda x: x.name)
alt_cluster_1, alt_cluster_2 = clusters
self.assertEqual(cluster_1.location_id, alt_cluster_1.location_id)
self.assertEqual(alt_cluster_1.state, enums.Cluster.State.READY)
self.assertEqual(cluster_1.serve_nodes, alt_cluster_1.serve_nodes)
self.assertEqual(cluster_1.default_storage_type,
alt_cluster_1.default_storage_type)
self.assertEqual(cluster_2.location_id, alt_cluster_2.location_id)
self.assertEqual(alt_cluster_2.state, enums.Cluster.State.READY)
self.assertEqual(cluster_2.serve_nodes, alt_cluster_2.serve_nodes)
self.assertEqual(cluster_2.default_storage_type,
alt_cluster_2.default_storage_type)
# Test list clusters in project via 'client.list_clusters'
clusters, failed_locations = Config.CLIENT.list_clusters()
self.assertFalse(failed_locations)
found = set([cluster.name for cluster in clusters])
self.assertTrue({alt_cluster_1.name,
alt_cluster_2.name,
Config.CLUSTER.name}.issubset(found))
temp_table_id = 'test-get-cluster-states'
temp_table = instance.table(temp_table_id)
> temp_table.create()
tests/system.py:280:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/bigtable/table.py:218: in create
table=table, initial_splits=splits)
google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py:327: in create_table
request, retry=retry, timeout=timeout, metadata=metadata)
../api_core/google/api_core/gapic_v1/method.py:139: in __call__
return wrapped_func(*args, **kwargs)
../api_core/google/api_core/retry.py:260: in retry_wrapped_func
on_error=on_error,
../api_core/google/api_core/retry.py:177: in retry_target
return target()
../api_core/google/api_core/timeout.py:206: in func_with_timeout
return func(*args, **kwargs)
../api_core/google/api_core/grpc_helpers.py:61: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = DeadlineExceeded('Deadline Exceeded',)
from_value = <_Rendezvous of RPC that terminated with:
status = StatusCode.DEADLINE_EXCEED...all.cc","file_line":1099,"grpc_message":"Deadline Exceeded","grpc_status":4}"
>
def raise_from(value, from_value):
> raise value
E DeadlineExceeded: 504 Deadline Exceeded
../.nox/sys-2-7/lib/python2.7/site-packages/six.py:737: DeadlineExceeded
``` | process | bigtable test create instance w two clusters flakes with deadline exceeded cc sduskis vikas jamdar from python testinstanceadminapi test create instance w two clusters self def test create instance w two clusters self from google cloud bigtable import enums from google cloud bigtable table import clusterstate production enums instance type production alt instance id dif unique resource id instance config client instance alt instance id instance type production labels labels alt cluster id alt instance id alt cluster id alt instance id location id us f storage type enums storagetype hdd cluster instance cluster alt cluster id location id location id serve nodes serve nodes default storage type storage type cluster instance cluster alt cluster id location id location id serve nodes serve nodes default storage type storage type operation instance create clusters we want to make sure the operation completes operation result timeout make sure this instance gets deleted after the test case self instances to delete append instance create a new instance instance and make sure it is the same instance alt config client instance alt instance id instance alt reload self assertequal instance instance alt self assertequal instance display name instance alt display name self assertequal instance type instance alt type clusters failed locations instance alt list clusters self assertequal failed locations clusters sort key lambda x x name alt cluster alt cluster clusters self assertequal cluster location id alt cluster location id self assertequal alt cluster state enums cluster state ready self assertequal cluster serve nodes alt cluster serve nodes self assertequal cluster default storage type alt cluster default storage type self assertequal cluster location id alt cluster location id self assertequal alt cluster state enums cluster state ready self assertequal cluster serve nodes alt cluster serve nodes self assertequal cluster default storage type alt cluster default storage type test list clusters in project via client list clusters clusters failed locations config client list clusters self assertfalse failed locations found set self asserttrue alt cluster name alt cluster name config cluster name issubset found temp table id test get cluster states temp table instance table temp table id temp table create tests system py google cloud bigtable table py in create table table initial splits splits google cloud bigtable admin gapic bigtable table admin client py in create table request retry retry timeout timeout metadata metadata api core google api core gapic method py in call return wrapped func args kwargs api core google api core retry py in retry wrapped func on error on error api core google api core retry py in retry target return target api core google api core timeout py in func with timeout return func args kwargs api core google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value deadlineexceeded deadline exceeded from value rendezvous of rpc that terminated with status statuscode deadline exceed all cc file line grpc message deadline exceeded grpc status def raise from value from value raise value e deadlineexceeded deadline exceeded nox sys lib site packages six py deadlineexceeded | 1 |
2,191 | 5,037,520,038 | IssuesEvent | 2016-12-17 18:23:11 | paulkornikov/Pragonas | https://api.github.com/repos/paulkornikov/Pragonas | closed | Perf sur trace détail | a-enhancement processus t-performance workload III | Créer une requête pour ne récupérer que le détail d'une trace et pas toutes les traces puis sélection d'une.
Modifier le web api en requêtant directement l'id dans le remote sur le client. | 1.0 | Perf sur trace détail - Créer une requête pour ne récupérer que le détail d'une trace et pas toutes les traces puis sélection d'une.
Modifier le web api en requêtant directement l'id dans le remote sur le client. | process | perf sur trace détail créer une requête pour ne récupérer que le détail d une trace et pas toutes les traces puis sélection d une modifier le web api en requêtant directement l id dans le remote sur le client | 1 |
86,174 | 24,776,969,887 | IssuesEvent | 2022-10-23 21:06:00 | microsoft/azure-pipelines-tasks | https://api.github.com/repos/microsoft/azure-pipelines-tasks | closed | MSBuildV1 task delays to find target solutions as part of *.sln pattern | bug Area: ABTT stale Task: MSBuild | ## Required Information
Entering this information will route you directly to the right team and expedite traction.
**Question, Bug, or Feature?**
*Type*: Bug
**Enter Task Name**: MSBuildV1
## Environment
- Server - Azure Pipelines or TFS on-premises?
- If using Azure Pipelines, provide the account name, team project name, build definition name/build number: [Build Link](https://dev.azure.com/lhediscovery/BET/_build/results?buildId=50985&view=results)
- Agent - Hosted or Private: Private Agent
- If using private agent, provide the OS of the machine running the agent and the agent version: AzAgent v2.193.0 running on Win2016
## Issue Description
MSBuild task takes a long time to build multiple solutions when using the '*.sln' glob pattern instead of targeting each solution in the folder.
### Task logs
[buildid-50985-raw-logs.zip](https://github.com/microsoft/azure-pipelines-tasks/files/7666642/buildid-50985-raw-logs.zip)
Log snippet that is most helpful:
```
2021-12-07T01:10:50.0886164Z ##[debug]Entering Get-SolutionFiles.
2021-12-07T01:10:50.1717793Z ##[debug] Solution: 'D:\agent_A\_work\423\s\*.sln'
2021-12-07T01:10:50.2697490Z ##[debug]Entering Find-VstsFiles.
2021-12-07T01:10:50.3544199Z ##[debug] LegacyPattern: 'D:\agent_A\_work\423\s\*.sln'
2021-12-07T01:10:50.5494214Z ##[debug]Entering Get-MatchingItems.
2021-12-07T01:10:50.6232044Z ##[debug] IncludePatterns: 'D:\agent_A\_work\423\s\*.sln'
2021-12-07T01:10:50.7008127Z ##[debug] ExcludePatterns: ''
2021-12-07T01:10:50.7787309Z ##[debug] IncludeFiles: 'True'
2021-12-07T01:10:50.8487212Z ##[debug] IncludeDirectories: 'False'
2021-12-07T01:10:50.9216358Z ##[debug] Force: 'False'
2021-12-07T01:21:01.4457554Z ##[debug]Paths: D:\agent_A\_work\423\s\
2021-12-07T01:21:01.5200358Z ##[debug] Core.DataProcessing.FilesMoverWorker.sln
2021-12-07T01:21:01.5897387Z ##[debug] Core.DataProcessing.NuixWorker.sln
2021-12-07T01:21:01.6609612Z ##[debug] Core.Exports.ExportPrepWorker.sln
2021-12-07T01:21:01.7317972Z ##[debug] Core.Exports.ExportWorker.sln
2021-12-07T01:21:01.8009118Z ##[debug] Core.Exports.RedemptionConverterService.sln
2021-12-07T01:21:01.8717068Z ##[debug] Core.Exports.RelativityImportWorker.sln
2021-12-07T01:21:01.9407503Z ##[debug] Core.Processing.DAL.SQLCLR.sln
2021-12-07T01:21:02.0119661Z ##[debug] Core.Processing.sln
2021-12-07T01:21:02.0826639Z ##[debug] Core.Reporting.sln
2021-12-07T01:21:02.1522239Z ##[debug] Core.Reporting.TestHarness.sln
2021-12-07T01:21:02.2234138Z ##[debug] Core.Tests.Aggregated.sln
2021-12-07T01:21:02.2924392Z ##[debug] Core.Workflow.JobsAPI.sln
2021-12-07T01:21:02.3638487Z ##[debug] Core.Workflow.QueueManager.sln
2021-12-07T01:21:02.4469705Z ##[debug]Leaving Get-MatchingItems.
2021-12-07T01:21:02.5341306Z ##[debug]Total found: 13
2021-12-07T01:21:02.6177091Z ##[debug]Leaving Find-VstsFiles.
2021-12-07T01:21:02.7060694Z ##[debug]Leaving Get-SolutionFiles.
```
### Error logs
No Error logs as build does succeed eventually but takes a long time to start the task.
| 1.0 | MSBuildV1 task delays to find target solutions as part of *.sln pattern - ## Required Information
Entering this information will route you directly to the right team and expedite traction.
**Question, Bug, or Feature?**
*Type*: Bug
**Enter Task Name**: MSBuildV1
## Environment
- Server - Azure Pipelines or TFS on-premises?
- If using Azure Pipelines, provide the account name, team project name, build definition name/build number: [Build Link](https://dev.azure.com/lhediscovery/BET/_build/results?buildId=50985&view=results)
- Agent - Hosted or Private: Private Agent
- If using private agent, provide the OS of the machine running the agent and the agent version: AzAgent v2.193.0 running on Win2016
## Issue Description
MSBuild task takes a long time to build multiple solutions when using the '*.sln' glob pattern instead of targeting each solution in the folder.
### Task logs
[buildid-50985-raw-logs.zip](https://github.com/microsoft/azure-pipelines-tasks/files/7666642/buildid-50985-raw-logs.zip)
Log snippet that is most helpful:
```
2021-12-07T01:10:50.0886164Z ##[debug]Entering Get-SolutionFiles.
2021-12-07T01:10:50.1717793Z ##[debug] Solution: 'D:\agent_A\_work\423\s\*.sln'
2021-12-07T01:10:50.2697490Z ##[debug]Entering Find-VstsFiles.
2021-12-07T01:10:50.3544199Z ##[debug] LegacyPattern: 'D:\agent_A\_work\423\s\*.sln'
2021-12-07T01:10:50.5494214Z ##[debug]Entering Get-MatchingItems.
2021-12-07T01:10:50.6232044Z ##[debug] IncludePatterns: 'D:\agent_A\_work\423\s\*.sln'
2021-12-07T01:10:50.7008127Z ##[debug] ExcludePatterns: ''
2021-12-07T01:10:50.7787309Z ##[debug] IncludeFiles: 'True'
2021-12-07T01:10:50.8487212Z ##[debug] IncludeDirectories: 'False'
2021-12-07T01:10:50.9216358Z ##[debug] Force: 'False'
2021-12-07T01:21:01.4457554Z ##[debug]Paths: D:\agent_A\_work\423\s\
2021-12-07T01:21:01.5200358Z ##[debug] Core.DataProcessing.FilesMoverWorker.sln
2021-12-07T01:21:01.5897387Z ##[debug] Core.DataProcessing.NuixWorker.sln
2021-12-07T01:21:01.6609612Z ##[debug] Core.Exports.ExportPrepWorker.sln
2021-12-07T01:21:01.7317972Z ##[debug] Core.Exports.ExportWorker.sln
2021-12-07T01:21:01.8009118Z ##[debug] Core.Exports.RedemptionConverterService.sln
2021-12-07T01:21:01.8717068Z ##[debug] Core.Exports.RelativityImportWorker.sln
2021-12-07T01:21:01.9407503Z ##[debug] Core.Processing.DAL.SQLCLR.sln
2021-12-07T01:21:02.0119661Z ##[debug] Core.Processing.sln
2021-12-07T01:21:02.0826639Z ##[debug] Core.Reporting.sln
2021-12-07T01:21:02.1522239Z ##[debug] Core.Reporting.TestHarness.sln
2021-12-07T01:21:02.2234138Z ##[debug] Core.Tests.Aggregated.sln
2021-12-07T01:21:02.2924392Z ##[debug] Core.Workflow.JobsAPI.sln
2021-12-07T01:21:02.3638487Z ##[debug] Core.Workflow.QueueManager.sln
2021-12-07T01:21:02.4469705Z ##[debug]Leaving Get-MatchingItems.
2021-12-07T01:21:02.5341306Z ##[debug]Total found: 13
2021-12-07T01:21:02.6177091Z ##[debug]Leaving Find-VstsFiles.
2021-12-07T01:21:02.7060694Z ##[debug]Leaving Get-SolutionFiles.
```
### Error logs
No Error logs as build does succeed eventually but takes a long time to start the task.
| non_process | task delays to find target solutions as part of sln pattern required information entering this information will route you directly to the right team and expedite traction question bug or feature type bug enter task name environment server azure pipelines or tfs on premises if using azure pipelines provide the account name team project name build definition name build number agent hosted or private private agent if using private agent provide the os of the machine running the agent and the agent version azagent running on issue description msbuild task takes a long time to build multiple solutions when using the sln glob pattern instead of targeting each solution in the folder task logs log snippet that is most helpful entering get solutionfiles solution d agent a work s sln entering find vstsfiles legacypattern d agent a work s sln entering get matchingitems includepatterns d agent a work s sln excludepatterns includefiles true includedirectories false force false paths d agent a work s core dataprocessing filesmoverworker sln core dataprocessing nuixworker sln core exports exportprepworker sln core exports exportworker sln core exports redemptionconverterservice sln core exports relativityimportworker sln core processing dal sqlclr sln core processing sln core reporting sln core reporting testharness sln core tests aggregated sln core workflow jobsapi sln core workflow queuemanager sln leaving get matchingitems total found leaving find vstsfiles leaving get solutionfiles error logs no error logs as build does succeed eventually but takes a long time to start the task | 0 |
20,981 | 27,844,686,947 | IssuesEvent | 2023-03-20 14:49:25 | camunda/issues | https://api.github.com/repos/camunda/issues | opened | Process Instance Version Migration | component:operate component:zeebe component:zeebe-process-automation public feature-parity potential:8.3 | ### Value Proposition Statement
Migrate running Process Instances between different versions of process definitions.
### User Problem
Migration itself:
- Our Operators have a new version of a workflow and want to move all the running instances from the old workflows to this new version because the other workflow versions are either outdated or have an error.
- Currently, when I deploy a new version of a process definition and want to run it in the new version, I need to cancel the old instance and recreate it in the new version of the process definition with the same context (probably via start process instance anywhere).
- If a process instance(s) has an incident or a message does not arrive, and if it requires an update in the process definition, then they deploy a new version of the process definition. In this case, they need to migrate all the process instances which are stuck in the previous version to this new version.
Around migration (based on Camunda 7):
- [After migration is done, the numbers won't sum up for all the flow nodes - e.g. 4 activities where executed on V1, than PI migrated and continued on V2 - we should communicate this in UI](https://jira.camunda.com/browse/SUPPORT-13873)
- [I want to see the previously completed activities in the process instance after migration](https://jira.camunda.com/browse/CAM-14466)
- https://jira.camunda.com/browse/SUPPORT-13237
### User Stories
- As an Operator, I can migrate all running process instances from one version to another.
- The target version can be higher than the source version
- The target version can be lower that the source version
- As an Operator, I can migrate a chosen set of running process instances from one version to a different version
- As an Operator, I can clearly map and see the migration plan - what flow node instances will be migrated and where
- As an Operator, I can add variables to migrated instances
- As an Operator, I can see the migration in the history log and a link between source and target instances
- As an Operator, I can migrate instances to the different version via Operate UI and API
### Implementation Notes
#### Requirements
1. Functional Requirements
- Select origin and target workflow and version
- Describe migration instructions
- Add variables to migrated instances
- Apply the migration
- Confirmation of the operation
- Providing an overview before and after the operation
- Do migration of multiple instances
- There should be a link to indicate source instance
- Modification should be indicated in the history log
2. Non-functional Requirements
- Scalability: Should be applicable for a big number of instances
#### Assumptions
- The same operations can be done via UI and API - migrating running process instances between versions
- One-to-one relationship when migrating
- Vast majority is adding new tasks, removing is not common.
#### Open questions
- How does users expect to see the migrated instances?
- In the new (target) definition
- In the old (source) definition
- Should the processInstanceKey change?
- Can we migrate between different type of activities?
### Validation Criteria
- Number of migrated instances / api calls to the migration endpoint
- At least 3 C8 customers are aware of the feature and have adapted it.
### Links
- https://docs.camunda.org/manual/7.16/user-guide/process-engine/process-instance-migration/
- [Iterations for Process Instance Version Migration](https://miro.com/app/board/uXjVPOH18_0=/)
- [PM Summary of the PI Version Migration](https://docs.google.com/document/d/10WHDG2Zv_DYVoMPVulAC_Ns8cOaFwBv18e6aLtZwpPU/edit)
### Breakdown
#### Discovery phase ##
1. User journey
- User selects the origin workflow and its version.
- User filters the instance(s) that should be migrated.
- User selects instance(s) for migration.
- User selects the targeted workflow and its version.
- User describes how to migrate.
- User confirms to apply migration of workflow instances.
2. Motivation
- Customers try to have only 1 active version of definition
- The code is always running in the latest version to reduce complexity of the code to support older versions
- When having Long-running processes (for months or years), it’s important to be able to introduce changes to a process definition and running instances. Let’s say, we have added a new sales channel or we need to comply with different regulatory requirements, so the process has to change. With that change, we want to our running instances to be on the new version of the diagram, to reflect the new business situation and comply with regulations.
- I want to change process definition due to the bug or business improvement. After deploying the new version, I would like to migrate my running instances to unblock them.
- Process instance migration for operators will ensure that all process instances are running on the correct version of the process.
3. Use cases:
- Migrating all the running instances from one version to another of a process definition
- Upgrading running instances to fix a bug in the old workflow version
- Downgrading running instances in a previous version
- Business changes generate new version
- When I deploy a new version of a process definition, I want to migrate all the process instances to the newest version
- Migrating instances into another workflow
- Migrating a set of running instances into a specific workflow version
- A/B testing of a workflow
- Need to migrate multiple workflows as one workflow has many child processes
4. Pain points in Camunda 7 Cockpit
- Cockpit UI is overwhelming
- Too many arrows
- Summary of the migration plan is overwhelming - most of the names will be repeated for source and target
- Make it foldable to extend/collapse if needed
- Need to see only the activities that they changes manually
- Options that users do not understand
- **"Link diagrams navigation"**
- 2/3 options do not have explanation
- No info that existing variables will be kept
- Good to see numbers of instances in every activity
- No easy way to confim if the IDs are correct - I need to have name etc.
- **[Define mapping screen]:**
- The mapping is hard to digest, the difference between 2 diagrams is not clear. Adding the layer of migration plan (green arrows) creates information overload for the user.
- "Link diagrams navigation" naming and meaning are not clear and not-known (even by a very experienced users)
- Confusing that not all activities have matching arrows
- [positive] Good to see the number of running instances
- **[Set variables screen]:**
- Not enough feedback: not clear that all variables will be kept.
- **[Select instances screen]**
- ID's and business key's do not provide information because the most important is to know the process definition key
- In case the list may includes thousands of items --> the screen get overwhelming
- **[Confirm screen]**
- [positive] Short explanation below the options help to understand what the feature does
- The explanation is given only to "Asynchronous" option and is missing (but expected) for "Skip Custom Listeners" and "Skip IO mapping"
- The readability of the summary is very low as it has gaps between information bits
- Migration plan has low readability: when source and target activities names are identical --> the screen does not deliver value to the user + lots of space between the lists
#### Define phase ##
Design Planning
* Reviewed by design: August 2022, 3 Jan 2023
* Designer assigned: Yes
* Assignee: @gastonpillet01
* [Design Brief](https://docs.google.com/document/d/1GT0a80wBexvXLCvDpSWxSDXIKs90oOr46LErpLtM1Os/edit?userstoinvite=johan.welgemoed@camunda.com&actionButton=1#) https://github.com/camunda/product-design/issues/75 Delivered Jan 17, 2023
* [Research Brief](https://docs.google.com/document/d/1k1sZLy7sD6Rw8endR4IzKka9EhmlQvkn2-kBkM6kfUk/edit#)
Design Deliverables (WIP)
- [Low-Fidelity Wireframes](https://www.figma.com/file/pb1vjdcPrcizCWN8HX2VFf/PVM-flows?node-id=0%3A1&t=3uIjuSDlbnscKVt7-1) - https://github.com/camunda/product-design/issues/53 - Expected: Feb 15, 2023
- [Wireframes](https://github.com/camunda/product-design/issues/54) (Expected delivery date ??) - Delivered: ??
- [Prototype](https://github.com/camunda/product-design/issues/51) (Expected delivery date ??) - Delivered: ??
- [Specifications](https://github.com/camunda/product-design/issues/52) (Expected delivery date ??) - Delivered: ??
- Handover Recording
Documentation Planning
<!-- Complex changes must be reviewed during the Define phase by the DRI of Documentation or technical writer. -->
<!-- Briefly describe the anticipated impact to documentation. -->
<!-- Example: "Creates structural changes in docs as UX is reworked." _Add docs reviewer to Epic for feedback._ -->
Risk Management <!-- add link to risk management issue -->
* Risk Class: <!-- e.g. very low | low | medium | high | very high -->
* Risk Treatment: <!-- e.g. avoid | mitigate | transfer | accept -->
#### Implement phase ##
#### Validate phase ##
### Links
#### Internal docs about PI Version Migration
- [Research](https://miro.com/app/board/uXjVOnwvLcs=/)
- [Interviews summary ](https://github.com/camunda/users-feedback/issues/116)
- [Research brief](https://docs.google.com/document/d/1k1sZLy7sD6Rw8endR4IzKka9EhmlQvkn2-kBkM6kfUk/edit)
- [Participants](https://docs.google.com/spreadsheets/d/1uy4SeNfs6XfVYKBilJND7M7kknHj9k_f7Kn4Q9UTYTs/edit#gid=0)
- [Interview questions](https://docs.google.com/presentation/d/14Sjsl-wW1wh-4TwIl-oWGPOkL__ShamACD9pasbVrQU/edit#slide=id.g72f03a6899_0_108)
- [Migration/Modification research results](https://miro.com/app/board/o9J_kqHjQUE=/)
- [Version migration summary](https://docs.google.com/presentation/d/1grPIqd_36DmOWwiZBQZdcm2gSsDrsbpf4Sy4Z5piJ0E/edit#slide=id.g72f03a6899_0_108)
- [Version Migration Survey](https://docs.google.com/spreadsheets/d/1CDPAqYpxa_jDTqfrBYvz8jPc23chReMQchk-bSDy_Sk/edit#gid=729869961)
- [Customer profiles](https://docs.google.com/document/d/1ckJiDl04Ve96Z1XW-98cBpFOh87Knm2xPeqRjgWRdII/edit#heading=h.8jsflevyw1hp)
- [User research](https://drive.google.com/drive/folders/1PhEMuAWWRhFbLgrlwITf_tGS8Ish10Xe)
| 1.0 | Process Instance Version Migration - ### Value Proposition Statement
Migrate running Process Instances between different versions of process definitions.
### User Problem
Migration itself:
- Our Operators have a new version of a workflow and want to move all the running instances from the old workflows to this new version because the other workflow versions are either outdated or have an error.
- Currently, when I deploy a new version of a process definition and want to run it in the new version, I need to cancel the old instance and recreate it in the new version of the process definition with the same context (probably via start process instance anywhere).
- If a process instance(s) has an incident or a message does not arrive, and if it requires an update in the process definition, then they deploy a new version of the process definition. In this case, they need to migrate all the process instances which are stuck in the previous version to this new version.
Around migration (based on Camunda 7):
- [After migration is done, the numbers won't sum up for all the flow nodes - e.g. 4 activities where executed on V1, than PI migrated and continued on V2 - we should communicate this in UI](https://jira.camunda.com/browse/SUPPORT-13873)
- [I want to see the previously completed activities in the process instance after migration](https://jira.camunda.com/browse/CAM-14466)
- https://jira.camunda.com/browse/SUPPORT-13237
### User Stories
- As an Operator, I can migrate all running process instances from one version to another.
- The target version can be higher than the source version
- The target version can be lower that the source version
- As an Operator, I can migrate a chosen set of running process instances from one version to a different version
- As an Operator, I can clearly map and see the migration plan - what flow node instances will be migrated and where
- As an Operator, I can add variables to migrated instances
- As an Operator, I can see the migration in the history log and a link between source and target instances
- As an Operator, I can migrate instances to the different version via Operate UI and API
### Implementation Notes
#### Requirements
1. Functional Requirements
- Select origin and target workflow and version
- Describe migration instructions
- Add variables to migrated instances
- Apply the migration
- Confirmation of the operation
- Providing an overview before and after the operation
- Do migration of multiple instances
- There should be a link to indicate source instance
- Modification should be indicated in the history log
2. Non-functional Requirements
- Scalability: Should be applicable for a big number of instances
#### Assumptions
- The same operations can be done via UI and API - migrating running process instances between versions
- One-to-one relationship when migrating
- Vast majority is adding new tasks, removing is not common.
#### Open questions
- How does users expect to see the migrated instances?
- In the new (target) definition
- In the old (source) definition
- Should the processInstanceKey change?
- Can we migrate between different type of activities?
### Validation Criteria
- Number of migrated instances / api calls to the migration endpoint
- At least 3 C8 customers are aware of the feature and have adapted it.
### Links
- https://docs.camunda.org/manual/7.16/user-guide/process-engine/process-instance-migration/
- [Iterations for Process Instance Version Migration](https://miro.com/app/board/uXjVPOH18_0=/)
- [PM Summary of the PI Version Migration](https://docs.google.com/document/d/10WHDG2Zv_DYVoMPVulAC_Ns8cOaFwBv18e6aLtZwpPU/edit)
### Breakdown
#### Discovery phase ##
1. User journey
- User selects the origin workflow and its version.
- User filters the instance(s) that should be migrated.
- User selects instance(s) for migration.
- User selects the targeted workflow and its version.
- User describes how to migrate.
- User confirms to apply migration of workflow instances.
2. Motivation
- Customers try to have only 1 active version of definition
- The code is always running in the latest version to reduce complexity of the code to support older versions
- When having Long-running processes (for months or years), it’s important to be able to introduce changes to a process definition and running instances. Let’s say, we have added a new sales channel or we need to comply with different regulatory requirements, so the process has to change. With that change, we want to our running instances to be on the new version of the diagram, to reflect the new business situation and comply with regulations.
- I want to change process definition due to the bug or business improvement. After deploying the new version, I would like to migrate my running instances to unblock them.
- Process instance migration for operators will ensure that all process instances are running on the correct version of the process.
3. Use cases:
- Migrating all the running instances from one version to another of a process definition
- Upgrading running instances to fix a bug in the old workflow version
- Downgrading running instances in a previous version
- Business changes generate new version
- When I deploy a new version of a process definition, I want to migrate all the process instances to the newest version
- Migrating instances into another workflow
- Migrating a set of running instances into a specific workflow version
- A/B testing of a workflow
- Need to migrate multiple workflows as one workflow has many child processes
4. Pain points in Camunda 7 Cockpit
- Cockpit UI is overwhelming
- Too many arrows
- Summary of the migration plan is overwhelming - most of the names will be repeated for source and target
- Make it foldable to extend/collapse if needed
- Need to see only the activities that they changes manually
- Options that users do not understand
- **"Link diagrams navigation"**
- 2/3 options do not have explanation
- No info that existing variables will be kept
- Good to see numbers of instances in every activity
- No easy way to confim if the IDs are correct - I need to have name etc.
- **[Define mapping screen]:**
- The mapping is hard to digest, the difference between 2 diagrams is not clear. Adding the layer of migration plan (green arrows) creates information overload for the user.
- "Link diagrams navigation" naming and meaning are not clear and not-known (even by a very experienced users)
- Confusing that not all activities have matching arrows
- [positive] Good to see the number of running instances
- **[Set variables screen]:**
- Not enough feedback: not clear that all variables will be kept.
- **[Select instances screen]**
- ID's and business key's do not provide information because the most important is to know the process definition key
- In case the list may includes thousands of items --> the screen get overwhelming
- **[Confirm screen]**
- [positive] Short explanation below the options help to understand what the feature does
- The explanation is given only to "Asynchronous" option and is missing (but expected) for "Skip Custom Listeners" and "Skip IO mapping"
- The readability of the summary is very low as it has gaps between information bits
- Migration plan has low readability: when source and target activities names are identical --> the screen does not deliver value to the user + lots of space between the lists
#### Define phase ##
Design Planning
* Reviewed by design: August 2022, 3 Jan 2023
* Designer assigned: Yes
* Assignee: @gastonpillet01
* [Design Brief](https://docs.google.com/document/d/1GT0a80wBexvXLCvDpSWxSDXIKs90oOr46LErpLtM1Os/edit?userstoinvite=johan.welgemoed@camunda.com&actionButton=1#) https://github.com/camunda/product-design/issues/75 Delivered Jan 17, 2023
* [Research Brief](https://docs.google.com/document/d/1k1sZLy7sD6Rw8endR4IzKka9EhmlQvkn2-kBkM6kfUk/edit#)
Design Deliverables (WIP)
- [Low-Fidelity Wireframes](https://www.figma.com/file/pb1vjdcPrcizCWN8HX2VFf/PVM-flows?node-id=0%3A1&t=3uIjuSDlbnscKVt7-1) - https://github.com/camunda/product-design/issues/53 - Expected: Feb 15, 2023
- [Wireframes](https://github.com/camunda/product-design/issues/54) (Expected delivery date ??) - Delivered: ??
- [Prototype](https://github.com/camunda/product-design/issues/51) (Expected delivery date ??) - Delivered: ??
- [Specifications](https://github.com/camunda/product-design/issues/52) (Expected delivery date ??) - Delivered: ??
- Handover Recording
Documentation Planning
<!-- Complex changes must be reviewed during the Define phase by the DRI of Documentation or technical writer. -->
<!-- Briefly describe the anticipated impact to documentation. -->
<!-- Example: "Creates structural changes in docs as UX is reworked." _Add docs reviewer to Epic for feedback._ -->
Risk Management <!-- add link to risk management issue -->
* Risk Class: <!-- e.g. very low | low | medium | high | very high -->
* Risk Treatment: <!-- e.g. avoid | mitigate | transfer | accept -->
#### Implement phase ##
#### Validate phase ##
### Links
#### Internal docs about PI Version Migration
- [Research](https://miro.com/app/board/uXjVOnwvLcs=/)
- [Interviews summary ](https://github.com/camunda/users-feedback/issues/116)
- [Research brief](https://docs.google.com/document/d/1k1sZLy7sD6Rw8endR4IzKka9EhmlQvkn2-kBkM6kfUk/edit)
- [Participants](https://docs.google.com/spreadsheets/d/1uy4SeNfs6XfVYKBilJND7M7kknHj9k_f7Kn4Q9UTYTs/edit#gid=0)
- [Interview questions](https://docs.google.com/presentation/d/14Sjsl-wW1wh-4TwIl-oWGPOkL__ShamACD9pasbVrQU/edit#slide=id.g72f03a6899_0_108)
- [Migration/Modification research results](https://miro.com/app/board/o9J_kqHjQUE=/)
- [Version migration summary](https://docs.google.com/presentation/d/1grPIqd_36DmOWwiZBQZdcm2gSsDrsbpf4Sy4Z5piJ0E/edit#slide=id.g72f03a6899_0_108)
- [Version Migration Survey](https://docs.google.com/spreadsheets/d/1CDPAqYpxa_jDTqfrBYvz8jPc23chReMQchk-bSDy_Sk/edit#gid=729869961)
- [Customer profiles](https://docs.google.com/document/d/1ckJiDl04Ve96Z1XW-98cBpFOh87Knm2xPeqRjgWRdII/edit#heading=h.8jsflevyw1hp)
- [User research](https://drive.google.com/drive/folders/1PhEMuAWWRhFbLgrlwITf_tGS8Ish10Xe)
| process | process instance version migration value proposition statement migrate running process instances between different versions of process definitions user problem migration itself our operators have a new version of a workflow and want to move all the running instances from the old workflows to this new version because the other workflow versions are either outdated or have an error currently when i deploy a new version of a process definition and want to run it in the new version i need to cancel the old instance and recreate it in the new version of the process definition with the same context probably via start process instance anywhere if a process instance s has an incident or a message does not arrive and if it requires an update in the process definition then they deploy a new version of the process definition in this case they need to migrate all the process instances which are stuck in the previous version to this new version around migration based on camunda user stories as an operator i can migrate all running process instances from one version to another the target version can be higher than the source version the target version can be lower that the source version as an operator i can migrate a chosen set of running process instances from one version to a different version as an operator i can clearly map and see the migration plan what flow node instances will be migrated and where as an operator i can add variables to migrated instances as an operator i can see the migration in the history log and a link between source and target instances as an operator i can migrate instances to the different version via operate ui and api implementation notes requirements functional requirements select origin and target workflow and version describe migration instructions add variables to migrated instances apply the migration confirmation of the operation providing an overview before and after the operation do migration of multiple instances there should be a link to indicate source instance modification should be indicated in the history log non functional requirements scalability should be applicable for a big number of instances assumptions the same operations can be done via ui and api migrating running process instances between versions one to one relationship when migrating vast majority is adding new tasks removing is not common open questions how does users expect to see the migrated instances in the new target definition in the old source definition should the processinstancekey change can we migrate between different type of activities validation criteria number of migrated instances api calls to the migration endpoint at least customers are aware of the feature and have adapted it links breakdown discovery phase user journey user selects the origin workflow and its version user filters the instance s that should be migrated user selects instance s for migration user selects the targeted workflow and its version user describes how to migrate user confirms to apply migration of workflow instances motivation customers try to have only active version of definition the code is always running in the latest version to reduce complexity of the code to support older versions when having long running processes for months or years it’s important to be able to introduce changes to a process definition and running instances let’s say we have added a new sales channel or we need to comply with different regulatory requirements so the process has to change with that change we want to our running instances to be on the new version of the diagram to reflect the new business situation and comply with regulations i want to change process definition due to the bug or business improvement after deploying the new version i would like to migrate my running instances to unblock them process instance migration for operators will ensure that all process instances are running on the correct version of the process use cases migrating all the running instances from one version to another of a process definition upgrading running instances to fix a bug in the old workflow version downgrading running instances in a previous version business changes generate new version when i deploy a new version of a process definition i want to migrate all the process instances to the newest version migrating instances into another workflow migrating a set of running instances into a specific workflow version a b testing of a workflow need to migrate multiple workflows as one workflow has many child processes pain points in camunda cockpit cockpit ui is overwhelming too many arrows summary of the migration plan is overwhelming most of the names will be repeated for source and target make it foldable to extend collapse if needed need to see only the activities that they changes manually options that users do not understand link diagrams navigation options do not have explanation no info that existing variables will be kept good to see numbers of instances in every activity no easy way to confim if the ids are correct i need to have name etc the mapping is hard to digest the difference between diagrams is not clear adding the layer of migration plan green arrows creates information overload for the user link diagrams navigation naming and meaning are not clear and not known even by a very experienced users confusing that not all activities have matching arrows good to see the number of running instances not enough feedback not clear that all variables will be kept id s and business key s do not provide information because the most important is to know the process definition key in case the list may includes thousands of items the screen get overwhelming short explanation below the options help to understand what the feature does the explanation is given only to asynchronous option and is missing but expected for skip custom listeners and skip io mapping the readability of the summary is very low as it has gaps between information bits migration plan has low readability when source and target activities names are identical the screen does not deliver value to the user lots of space between the lists define phase design planning reviewed by design august jan designer assigned yes assignee delivered jan design deliverables wip expected feb expected delivery date delivered expected delivery date delivered expected delivery date delivered handover recording documentation planning risk management risk class risk treatment implement phase validate phase links internal docs about pi version migration | 1 |
3,917 | 6,840,610,920 | IssuesEvent | 2017-11-11 01:56:33 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | When joining in a way that returns columns with same name, columns are suffixed and missing metadata | Limitation Query Processor | e.g. `name` and `name_2`. I believe this is done automatically by Clojure JDBC.
- Since we don't know the source of `name_2` the correct Metadata doesn't get returned along with it.
- It would be better to give it a name like `<table_name>.name`.
See also #1447.
| 1.0 | When joining in a way that returns columns with same name, columns are suffixed and missing metadata - e.g. `name` and `name_2`. I believe this is done automatically by Clojure JDBC.
- Since we don't know the source of `name_2` the correct Metadata doesn't get returned along with it.
- It would be better to give it a name like `<table_name>.name`.
See also #1447.
| process | when joining in a way that returns columns with same name columns are suffixed and missing metadata e g name and name i believe this is done automatically by clojure jdbc since we don t know the source of name the correct metadata doesn t get returned along with it it would be better to give it a name like name see also | 1 |
15,151 | 18,907,831,440 | IssuesEvent | 2021-11-16 10:57:57 | sillsdev/silnlp | https://api.github.com/repos/sillsdev/silnlp | opened | Improve the error message when a parent model folder is missing. (Translate) | enhancement good first issue pipeline 6: infer pipeline 3: preprocess pipeline 4: train pipeline 5: test | When a parent model folder is missing then the config file for that parent folder cannot be found. That leads to an error like this:
```
FileNotFoundError: [Errno 2] No such file or directory: '/home/david/gutenberg/MT/experiments/BT-Swahili/en-swh-8/config.yml'
```
I'm running a child model where the folder name begins en-swh-8 It took me a while to realise that there wasn't a problem with finding the child model config file but that the whole parent model folder was missing.
A few improvements would be:
1) Indicate to the user that the child model folder and config file exist.
2) If a file is missing indicate which parts of the path are missing.
3) If we could indicate that 'line x of the config file <path to config file> refers to file <path to missing file> which can't be found.' That would be very helpful.
4) Prior to preprocessing we check that the required files exist before trying to preprocess.
Ensure that similar checks occur (those that are required) for each part of the pipeline.
| 1.0 | Improve the error message when a parent model folder is missing. (Translate) - When a parent model folder is missing then the config file for that parent folder cannot be found. That leads to an error like this:
```
FileNotFoundError: [Errno 2] No such file or directory: '/home/david/gutenberg/MT/experiments/BT-Swahili/en-swh-8/config.yml'
```
I'm running a child model where the folder name begins en-swh-8 It took me a while to realise that there wasn't a problem with finding the child model config file but that the whole parent model folder was missing.
A few improvements would be:
1) Indicate to the user that the child model folder and config file exist.
2) If a file is missing indicate which parts of the path are missing.
3) If we could indicate that 'line x of the config file <path to config file> refers to file <path to missing file> which can't be found.' That would be very helpful.
4) Prior to preprocessing we check that the required files exist before trying to preprocess.
Ensure that similar checks occur (those that are required) for each part of the pipeline.
| process | improve the error message when a parent model folder is missing translate when a parent model folder is missing then the config file for that parent folder cannot be found that leads to an error like this filenotfounderror no such file or directory home david gutenberg mt experiments bt swahili en swh config yml i m running a child model where the folder name begins en swh it took me a while to realise that there wasn t a problem with finding the child model config file but that the whole parent model folder was missing a few improvements would be indicate to the user that the child model folder and config file exist if a file is missing indicate which parts of the path are missing if we could indicate that line x of the config file refers to file which can t be found that would be very helpful prior to preprocessing we check that the required files exist before trying to preprocess ensure that similar checks occur those that are required for each part of the pipeline | 1 |
15,219 | 19,086,515,394 | IssuesEvent | 2021-11-29 07:01:10 | varabyte/kobweb | https://api.github.com/repos/varabyte/kobweb | closed | Break up WebModifiers.kt | process | The file is getting unwieldy.
Maybe something like
WebModifiers.color.kt
WebModifiers.size.kt
WebModifiers.input.kt
etc. | 1.0 | Break up WebModifiers.kt - The file is getting unwieldy.
Maybe something like
WebModifiers.color.kt
WebModifiers.size.kt
WebModifiers.input.kt
etc. | process | break up webmodifiers kt the file is getting unwieldy maybe something like webmodifiers color kt webmodifiers size kt webmodifiers input kt etc | 1 |
211,776 | 16,457,816,897 | IssuesEvent | 2021-05-21 14:44:55 | composer/composer | https://api.github.com/repos/composer/composer | closed | No lock file found message is not intuitive | Documentation | Couple of times, I got a question about "No lock file found" warning in Composer 2. Newbies that never or rarely worked with Composer (or simply do not use it day-to-day) are asking why should they use `composer update` over `composer install` when no lock file is present. Their questions are mainly due to situations when there are no vendors installed at all, so it is confusing to them to have to run `composer update` because there is nothing to be updated yet.
While I understand and always explain the reasoning behind the message and am aware of potential edge cases (`vendor/` exists, but no lock file etc), I feel that the warning message brings more confusion than it benefits. At the very least, it should not be displayed when both `vendor` and `composer.lock` are not present.
My `composer.json`:
Not relevant
Output of `composer diagnose`:
```
Checking platform settings: OK
Checking git settings: OK
Checking http connectivity to packagist: OK
Checking https connectivity to packagist: OK
Checking github.com oauth access: OK
Checking disk free space: OK
Checking pubkeys:
Tags Public Key Fingerprint: 57815BA2 7E54DC31 7ECC7CC5 573090D0 87719BA6 8F3BB723 4E5D42D0 84A14642
Dev Public Key Fingerprint: 4AC45767 E5EC2265 2F0C1167 CBBB8A2B 0C708369 153E328C AD90147D AFE50952
OK
Checking composer version: OK
Composer version: 2.0.11
PHP version: 8.0.2
PHP binary path: /usr/bin/php8.0
OpenSSL version: OpenSSL 1.1.1j 16 Feb 2021
cURL version: 7.68.0 libz 1.2.11 ssl OpenSSL/1.1.1j
zip: extension present, unzip present
```
When I run this command (without `vendor` folder `composer.lock` file present):
```
composer install
```
I get the following output:
```
No lock file found. Updating dependencies instead of installing from lock file. Use composer update over composer install if you do not have a lock file.
```
And I expected this to happen:
No warning message. | 1.0 | No lock file found message is not intuitive - Couple of times, I got a question about "No lock file found" warning in Composer 2. Newbies that never or rarely worked with Composer (or simply do not use it day-to-day) are asking why should they use `composer update` over `composer install` when no lock file is present. Their questions are mainly due to situations when there are no vendors installed at all, so it is confusing to them to have to run `composer update` because there is nothing to be updated yet.
While I understand and always explain the reasoning behind the message and am aware of potential edge cases (`vendor/` exists, but no lock file etc), I feel that the warning message brings more confusion than it benefits. At the very least, it should not be displayed when both `vendor` and `composer.lock` are not present.
My `composer.json`:
Not relevant
Output of `composer diagnose`:
```
Checking platform settings: OK
Checking git settings: OK
Checking http connectivity to packagist: OK
Checking https connectivity to packagist: OK
Checking github.com oauth access: OK
Checking disk free space: OK
Checking pubkeys:
Tags Public Key Fingerprint: 57815BA2 7E54DC31 7ECC7CC5 573090D0 87719BA6 8F3BB723 4E5D42D0 84A14642
Dev Public Key Fingerprint: 4AC45767 E5EC2265 2F0C1167 CBBB8A2B 0C708369 153E328C AD90147D AFE50952
OK
Checking composer version: OK
Composer version: 2.0.11
PHP version: 8.0.2
PHP binary path: /usr/bin/php8.0
OpenSSL version: OpenSSL 1.1.1j 16 Feb 2021
cURL version: 7.68.0 libz 1.2.11 ssl OpenSSL/1.1.1j
zip: extension present, unzip present
```
When I run this command (without `vendor` folder `composer.lock` file present):
```
composer install
```
I get the following output:
```
No lock file found. Updating dependencies instead of installing from lock file. Use composer update over composer install if you do not have a lock file.
```
And I expected this to happen:
No warning message. | non_process | no lock file found message is not intuitive couple of times i got a question about no lock file found warning in composer newbies that never or rarely worked with composer or simply do not use it day to day are asking why should they use composer update over composer install when no lock file is present their questions are mainly due to situations when there are no vendors installed at all so it is confusing to them to have to run composer update because there is nothing to be updated yet while i understand and always explain the reasoning behind the message and am aware of potential edge cases vendor exists but no lock file etc i feel that the warning message brings more confusion than it benefits at the very least it should not be displayed when both vendor and composer lock are not present my composer json not relevant output of composer diagnose checking platform settings ok checking git settings ok checking http connectivity to packagist ok checking https connectivity to packagist ok checking github com oauth access ok checking disk free space ok checking pubkeys tags public key fingerprint dev public key fingerprint ok checking composer version ok composer version php version php binary path usr bin openssl version openssl feb curl version libz ssl openssl zip extension present unzip present when i run this command without vendor folder composer lock file present composer install i get the following output no lock file found updating dependencies instead of installing from lock file use composer update over composer install if you do not have a lock file and i expected this to happen no warning message | 0 |
600,025 | 18,288,041,443 | IssuesEvent | 2021-10-05 12:33:48 | UCREL/pymusas | https://api.github.com/repos/UCREL/pymusas | opened | Python 3.10 | enhancement low priority | When Conda releases Python version 3.10 we shall test against version 3.10 of Python. | 1.0 | Python 3.10 - When Conda releases Python version 3.10 we shall test against version 3.10 of Python. | non_process | python when conda releases python version we shall test against version of python | 0 |
22,409 | 31,142,292,680 | IssuesEvent | 2023-08-16 01:44:48 | cypress-io/cypress | https://api.github.com/repos/cypress-io/cypress | closed | Flaky test: must call #initializeConfig before accessing config | OS: linux process: flaky test topic: flake ❄️ stage: flake stale | ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41402/workflows/68bc48c2-74e7-4588-97c3-fbc1b0bde495/jobs/1714672
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/server/test/integration/cypress_spec.js#L1661
### Analysis
<img width="886" alt="Screen Shot 2022-08-05 at 12 22 06 PM" src="https://user-images.githubusercontent.com/26726429/183146676-4d212cc9-9ccb-41a3-af71-3a8e0125ee85.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed | 1.0 | Flaky test: must call #initializeConfig before accessing config - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41402/workflows/68bc48c2-74e7-4588-97c3-fbc1b0bde495/jobs/1714672
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/server/test/integration/cypress_spec.js#L1661
### Analysis
<img width="886" alt="Screen Shot 2022-08-05 at 12 22 06 PM" src="https://user-images.githubusercontent.com/26726429/183146676-4d212cc9-9ccb-41a3-af71-3a8e0125ee85.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed | process | flaky test must call initializeconfig before accessing config link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed | 1 |
602,316 | 18,466,684,859 | IssuesEvent | 2021-10-17 01:54:50 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Can't unbuckle mob from chair after pulling | Type: Bug Priority: 2-Before Release Difficulty: 2-Medium | <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] -->
## Description
1. Spawn Urist and a chair
2. Drag Urist on a chair and pull it for some time
3. You are not able unbuckle Urist from chair anymore
<!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.-->
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
https://user-images.githubusercontent.com/6161335/118340147-96000a80-b523-11eb-9fd6-8a768b1bda8e.mp4
**Additional context**
<!-- Add any other context about the problem here. -->
| 1.0 | Can't unbuckle mob from chair after pulling - <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] -->
## Description
1. Spawn Urist and a chair
2. Drag Urist on a chair and pull it for some time
3. You are not able unbuckle Urist from chair anymore
<!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.-->
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
https://user-images.githubusercontent.com/6161335/118340147-96000a80-b523-11eb-9fd6-8a768b1bda8e.mp4
**Additional context**
<!-- Add any other context about the problem here. -->
| non_process | can t unbuckle mob from chair after pulling description spawn urist and a chair drag urist on a chair and pull it for some time you are not able unbuckle urist from chair anymore screenshots additional context | 0 |
136,071 | 18,722,296,364 | IssuesEvent | 2021-11-03 13:09:12 | KDWSS/dd-trace-java | https://api.github.com/repos/KDWSS/dd-trace-java | opened | CVE-2014-7810 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2014-7810 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-jasper-7.0.20.jar</b>, <b>tomcat-el-api-7.0.20.jar</b>, <b>tomcat-el-api-8.0.14.jar</b>, <b>tomcat-embed-jasper-7.0.37.jar</b></p></summary>
<p>
<details><summary><b>tomcat-jasper-7.0.20.jar</b></p></summary>
<p>Tomcats JSP Parser</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/org.apache.tomcat/tomcat-jasper/7.0.20/ac4d5c47607c4a497c96dc07ecab91e9f247b665/tomcat-jasper-7.0.20.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-jasper-7.0.20.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-el-api-7.0.20.jar</b></p></summary>
<p>Expression language package</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.tomcat/tomcat-el-api/7.0.20/8eda5589dddbde5e4c70f5ce5d3c9f59a8c1000/tomcat-el-api-7.0.20.jar</p>
<p>
Dependency Hierarchy:
- tomcat-jasper-7.0.20.jar (Root Library)
- :x: **tomcat-el-api-7.0.20.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-el-api-8.0.14.jar</b></p></summary>
<p>Expression language package</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/classloading/tomcat-testing/tomcat-testing.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.tomcat/tomcat-el-api/8.0.14/6f4ec4c951ac2f6e129366ef0ef8621933e80607/tomcat-el-api-8.0.14.jar</p>
<p>
Dependency Hierarchy:
- tomcat-catalina-8.0.14.jar (Root Library)
- tomcat-jsp-api-8.0.14.jar
- :x: **tomcat-el-api-8.0.14.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-jasper-7.0.37.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-jasper/7.0.37/aeb4fef6e68d8df415e2f7d57dfcdd6dc1a43539/tomcat-embed-jasper-7.0.37.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-jasper-7.0.37.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/KDWSS/dd-trace-java/commit/2819174635979a19573ec0ce8e3e2b63a3848079">2819174635979a19573ec0ce8e3e2b63a3848079</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Expression Language (EL) implementation in Apache Tomcat 6.x before 6.0.44, 7.x before 7.0.58, and 8.x before 8.0.16 does not properly consider the possibility of an accessible interface implemented by an inaccessible class, which allows attackers to bypass a SecurityManager protection mechanism via a web application that leverages use of incorrect privileges during EL evaluation.
<p>Publish Date: 2015-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7810>CVE-2014-7810</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-7810">https://nvd.nist.gov/vuln/detail/CVE-2014-7810</a></p>
<p>Release Date: 2015-06-07</p>
<p>Fix Resolution: 6.0.44,7.0.58,8.0.16</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat","packageName":"tomcat-jasper","packageVersion":"7.0.20","packageFilePaths":["/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat:tomcat-jasper:7.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"},{"packageType":"Java","groupId":"org.apache.tomcat","packageName":"tomcat-el-api","packageVersion":"7.0.20","packageFilePaths":["/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tomcat:tomcat-jasper:7.0.20;org.apache.tomcat:tomcat-el-api:7.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"},{"packageType":"Java","groupId":"org.apache.tomcat","packageName":"tomcat-el-api","packageVersion":"8.0.14","packageFilePaths":["/dd-java-agent/instrumentation/classloading/tomcat-testing/tomcat-testing.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tomcat:tomcat-catalina:8.0.14;org.apache.tomcat:tomcat-jsp-api:8.0.14;org.apache.tomcat:tomcat-el-api:8.0.14","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"},{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-jasper","packageVersion":"7.0.37","packageFilePaths":["/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat.embed:tomcat-embed-jasper:7.0.37","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-7810","vulnerabilityDetails":"The Expression Language (EL) implementation in Apache Tomcat 6.x before 6.0.44, 7.x before 7.0.58, and 8.x before 8.0.16 does not properly consider the possibility of an accessible interface implemented by an inaccessible class, which allows attackers to bypass a SecurityManager protection mechanism via a web application that leverages use of incorrect privileges during EL evaluation.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7810","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | True | CVE-2014-7810 (Medium) detected in multiple libraries - ## CVE-2014-7810 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-jasper-7.0.20.jar</b>, <b>tomcat-el-api-7.0.20.jar</b>, <b>tomcat-el-api-8.0.14.jar</b>, <b>tomcat-embed-jasper-7.0.37.jar</b></p></summary>
<p>
<details><summary><b>tomcat-jasper-7.0.20.jar</b></p></summary>
<p>Tomcats JSP Parser</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/org.apache.tomcat/tomcat-jasper/7.0.20/ac4d5c47607c4a497c96dc07ecab91e9f247b665/tomcat-jasper-7.0.20.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-jasper-7.0.20.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-el-api-7.0.20.jar</b></p></summary>
<p>Expression language package</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.tomcat/tomcat-el-api/7.0.20/8eda5589dddbde5e4c70f5ce5d3c9f59a8c1000/tomcat-el-api-7.0.20.jar</p>
<p>
Dependency Hierarchy:
- tomcat-jasper-7.0.20.jar (Root Library)
- :x: **tomcat-el-api-7.0.20.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-el-api-8.0.14.jar</b></p></summary>
<p>Expression language package</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/classloading/tomcat-testing/tomcat-testing.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.tomcat/tomcat-el-api/8.0.14/6f4ec4c951ac2f6e129366ef0ef8621933e80607/tomcat-el-api-8.0.14.jar</p>
<p>
Dependency Hierarchy:
- tomcat-catalina-8.0.14.jar (Root Library)
- tomcat-jsp-api-8.0.14.jar
- :x: **tomcat-el-api-8.0.14.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-jasper-7.0.37.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-jasper/7.0.37/aeb4fef6e68d8df415e2f7d57dfcdd6dc1a43539/tomcat-embed-jasper-7.0.37.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-jasper-7.0.37.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/KDWSS/dd-trace-java/commit/2819174635979a19573ec0ce8e3e2b63a3848079">2819174635979a19573ec0ce8e3e2b63a3848079</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Expression Language (EL) implementation in Apache Tomcat 6.x before 6.0.44, 7.x before 7.0.58, and 8.x before 8.0.16 does not properly consider the possibility of an accessible interface implemented by an inaccessible class, which allows attackers to bypass a SecurityManager protection mechanism via a web application that leverages use of incorrect privileges during EL evaluation.
<p>Publish Date: 2015-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7810>CVE-2014-7810</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-7810">https://nvd.nist.gov/vuln/detail/CVE-2014-7810</a></p>
<p>Release Date: 2015-06-07</p>
<p>Fix Resolution: 6.0.44,7.0.58,8.0.16</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat","packageName":"tomcat-jasper","packageVersion":"7.0.20","packageFilePaths":["/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat:tomcat-jasper:7.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"},{"packageType":"Java","groupId":"org.apache.tomcat","packageName":"tomcat-el-api","packageVersion":"7.0.20","packageFilePaths":["/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tomcat:tomcat-jasper:7.0.20;org.apache.tomcat:tomcat-el-api:7.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"},{"packageType":"Java","groupId":"org.apache.tomcat","packageName":"tomcat-el-api","packageVersion":"8.0.14","packageFilePaths":["/dd-java-agent/instrumentation/classloading/tomcat-testing/tomcat-testing.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tomcat:tomcat-catalina:8.0.14;org.apache.tomcat:tomcat-jsp-api:8.0.14;org.apache.tomcat:tomcat-el-api:8.0.14","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"},{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-jasper","packageVersion":"7.0.37","packageFilePaths":["/dd-java-agent/instrumentation/jsp-2.3/jsp-2.3.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat.embed:tomcat-embed-jasper:7.0.37","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.44,7.0.58,8.0.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-7810","vulnerabilityDetails":"The Expression Language (EL) implementation in Apache Tomcat 6.x before 6.0.44, 7.x before 7.0.58, and 8.x before 8.0.16 does not properly consider the possibility of an accessible interface implemented by an inaccessible class, which allows attackers to bypass a SecurityManager protection mechanism via a web application that leverages use of incorrect privileges during EL evaluation.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7810","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | non_process | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries tomcat jasper jar tomcat el api jar tomcat el api jar tomcat embed jasper jar tomcat jasper jar tomcats jsp parser path to dependency file dd trace java dd java agent instrumentation jsp jsp gradle path to vulnerable library caches modules files org apache tomcat tomcat jasper tomcat jasper jar dependency hierarchy x tomcat jasper jar vulnerable library tomcat el api jar expression language package path to dependency file dd trace java dd java agent instrumentation jsp jsp gradle path to vulnerable library home wss scanner gradle caches modules files org apache tomcat tomcat el api tomcat el api jar dependency hierarchy tomcat jasper jar root library x tomcat el api jar vulnerable library tomcat el api jar expression language package library home page a href path to dependency file dd trace java dd java agent instrumentation classloading tomcat testing tomcat testing gradle path to vulnerable library home wss scanner gradle caches modules files org apache tomcat tomcat el api tomcat el api jar dependency hierarchy tomcat catalina jar root library tomcat jsp api jar x tomcat el api jar vulnerable library tomcat embed jasper jar core tomcat implementation path to dependency file dd trace java dd java agent instrumentation jsp jsp gradle path to vulnerable library caches modules files org apache tomcat embed tomcat embed jasper tomcat embed jasper jar dependency hierarchy x tomcat embed jasper jar vulnerable library found in head commit a href found in base branch master vulnerability details the expression language el implementation in apache tomcat x before x before and x before does not properly consider the possibility of an accessible interface implemented by an inaccessible class which allows attackers to bypass a securitymanager protection mechanism via a web application that leverages use of incorrect privileges during el evaluation publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache tomcat tomcat jasper isminimumfixversionavailable true minimumfixversion packagetype java groupid org apache tomcat packagename tomcat el api packageversion packagefilepaths istransitivedependency true dependencytree org apache tomcat tomcat jasper org apache tomcat tomcat el api isminimumfixversionavailable true minimumfixversion packagetype java groupid org apache tomcat packagename tomcat el api packageversion packagefilepaths istransitivedependency true dependencytree org apache tomcat tomcat catalina org apache tomcat tomcat jsp api org apache tomcat tomcat el api isminimumfixversionavailable true minimumfixversion packagetype java groupid org apache tomcat embed packagename tomcat embed jasper packageversion packagefilepaths istransitivedependency false dependencytree org apache tomcat embed tomcat embed jasper isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the expression language el implementation in apache tomcat x before x before and x before does not properly consider the possibility of an accessible interface implemented by an inaccessible class which allows attackers to bypass a securitymanager protection mechanism via a web application that leverages use of incorrect privileges during el evaluation vulnerabilityurl | 0 |
5,876 | 2,796,615,715 | IssuesEvent | 2015-05-12 08:45:49 | mysociety/popolo-viewer-sinatra | https://api.github.com/repos/mysociety/popolo-viewer-sinatra | opened | Make the Country List page look nicer | 1 - Desired design | At the moment http://data.everypolitician.org/ is a separate homepage with a list of countries. Ideally we'll consolidate the sites into one (requires #120), but we'll always need _something_ to show which countries we have, and a non-zoomable map on its own isn't enough.
Unless or until we resolve the wider issue, it's probably worth doing something quite basic to make the http://data.everypolitician.org/ page itself slightly nicer. | 1.0 | Make the Country List page look nicer - At the moment http://data.everypolitician.org/ is a separate homepage with a list of countries. Ideally we'll consolidate the sites into one (requires #120), but we'll always need _something_ to show which countries we have, and a non-zoomable map on its own isn't enough.
Unless or until we resolve the wider issue, it's probably worth doing something quite basic to make the http://data.everypolitician.org/ page itself slightly nicer. | non_process | make the country list page look nicer at the moment is a separate homepage with a list of countries ideally we ll consolidate the sites into one requires but we ll always need something to show which countries we have and a non zoomable map on its own isn t enough unless or until we resolve the wider issue it s probably worth doing something quite basic to make the page itself slightly nicer | 0 |
11,401 | 14,235,750,911 | IssuesEvent | 2020-11-18 15:11:34 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | How to set the cwd when a genrule's cmd is executed | team-Rules-Server type: support / not a bug (process) untriaged | Hi,
My project depends on an external git repository and I add it to the WORKSPACE file using `new_git_repository`. This external git repository has a python module and I need to execute it with `python -m module_name` to generate some C++ source and header files for the main repository. I use genrule with a `cmd = "python -m module_name"`, but bazel outputs: No module named module_name. Then I use `cmd = "pwd"` to see where I am, it tells me I'm not in the root directory of the external git repository. How to change the cwd of the execution of the `cmd` to the root directory of the external git repository or how to run a python module in the external git repository?
### What operating system are you running Bazel on?
> Ubuntu 16.04
### What's the output of `bazel info release`?
> release 3.7.0
### Have you found anything relevant by searching the web?
> Google "bazel genrule change cwd" | 1.0 | How to set the cwd when a genrule's cmd is executed - Hi,
My project depends on an external git repository and I add it to the WORKSPACE file using `new_git_repository`. This external git repository has a python module and I need to execute it with `python -m module_name` to generate some C++ source and header files for the main repository. I use genrule with a `cmd = "python -m module_name"`, but bazel outputs: No module named module_name. Then I use `cmd = "pwd"` to see where I am, it tells me I'm not in the root directory of the external git repository. How to change the cwd of the execution of the `cmd` to the root directory of the external git repository or how to run a python module in the external git repository?
### What operating system are you running Bazel on?
> Ubuntu 16.04
### What's the output of `bazel info release`?
> release 3.7.0
### Have you found anything relevant by searching the web?
> Google "bazel genrule change cwd" | process | how to set the cwd when a genrule s cmd is executed hi my project depends on an external git repository and i add it to the workspace file using new git repository this external git repository has a python module and i need to execute it with python m module name to generate some c source and header files for the main repository i use genrule with a cmd python m module name but bazel outputs no module named module name then i use cmd pwd to see where i am it tells me i m not in the root directory of the external git repository how to change the cwd of the execution of the cmd to the root directory of the external git repository or how to run a python module in the external git repository what operating system are you running bazel on ubuntu what s the output of bazel info release release have you found anything relevant by searching the web google bazel genrule change cwd | 1 |
121,075 | 15,836,738,557 | IssuesEvent | 2021-04-06 19:46:12 | microsoft/WSL | https://api.github.com/repos/microsoft/WSL | closed | Calling `explorer.exe /directory` from within an Ubuntu instance opens the explorer but on an empty view and not on the specified directory | bydesign | Using Windows Insider 10.0.18932.1000 with Ubuntu 18.04, doing this from the Ubuntu command line
$ explorer.exe /etc
opens a File Explorer view which states 'This folder is empty"
While doing
$ cd /etc; explorer.exe .
Correctly open a File Explorer listing the directory content
They should be equivalent commands | 1.0 | Calling `explorer.exe /directory` from within an Ubuntu instance opens the explorer but on an empty view and not on the specified directory - Using Windows Insider 10.0.18932.1000 with Ubuntu 18.04, doing this from the Ubuntu command line
$ explorer.exe /etc
opens a File Explorer view which states 'This folder is empty"
While doing
$ cd /etc; explorer.exe .
Correctly open a File Explorer listing the directory content
They should be equivalent commands | non_process | calling explorer exe directory from within an ubuntu instance opens the explorer but on an empty view and not on the specified directory using windows insider with ubuntu doing this from the ubuntu command line explorer exe etc opens a file explorer view which states this folder is empty while doing cd etc explorer exe correctly open a file explorer listing the directory content they should be equivalent commands | 0 |
250,916 | 7,992,563,954 | IssuesEvent | 2018-07-20 02:17:35 | magda-io/magda | https://api.github.com/repos/magda-io/magda | closed | Launceston have changed the URL of their data.json page. | priority: showstopper | ### Problem description
It's now at "https://data-launceston.opendata.arcgis.com/data.json". We need to update this in our config. | 1.0 | Launceston have changed the URL of their data.json page. - ### Problem description
It's now at "https://data-launceston.opendata.arcgis.com/data.json". We need to update this in our config. | non_process | launceston have changed the url of their data json page problem description it s now at we need to update this in our config | 0 |
755,327 | 26,425,108,339 | IssuesEvent | 2023-01-14 03:49:11 | yogstation13/Yogstation | https://api.github.com/repos/yogstation13/Yogstation | closed | You can't light other people's cigarrettes with things that aren't lighters | Bug Issue - Confirmed Issue - Low priority | ## Reproduction:
Try to light someone else's cigarrette with a welding tool or laser
Cry when you just shoot them or hit them with it instead | 1.0 | You can't light other people's cigarrettes with things that aren't lighters - ## Reproduction:
Try to light someone else's cigarrette with a welding tool or laser
Cry when you just shoot them or hit them with it instead | non_process | you can t light other people s cigarrettes with things that aren t lighters reproduction try to light someone else s cigarrette with a welding tool or laser cry when you just shoot them or hit them with it instead | 0 |
38,868 | 8,556,473,593 | IssuesEvent | 2018-11-08 13:17:36 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.0] com_redirect bulk import | J4 Issue No Code Attached Yet | It is not possible to open the bulk import modal from the toolbar in com_redirect due to a javascript error
```
Uncaught TypeError: Cannot read property 'open' of null
at HTMLButtonElement.onclick (index.php?option=com_redirect&view=links:279)
``` | 1.0 | [4.0] com_redirect bulk import - It is not possible to open the bulk import modal from the toolbar in com_redirect due to a javascript error
```
Uncaught TypeError: Cannot read property 'open' of null
at HTMLButtonElement.onclick (index.php?option=com_redirect&view=links:279)
``` | non_process | com redirect bulk import it is not possible to open the bulk import modal from the toolbar in com redirect due to a javascript error uncaught typeerror cannot read property open of null at htmlbuttonelement onclick index php option com redirect view links | 0 |
14,597 | 17,703,567,913 | IssuesEvent | 2021-08-25 03:17:50 | tdwg/dwc | https://api.github.com/repos/tdwg/dwc | closed | Change term - coordinateUncertaintyInMeters | Term - change Class - Location non-normative Process - complete | ## Change term
* Submitter: https://github.com/RicardoOrtizG
* Justification (why is this change necessary?): For completeness
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/terms/#dwc:coordinateUncertaintynMeters
Proposed new attributes of the term:
* Term name (in lowerCamelCase): coordinateUncertaintyInMeters
* Organized in Class (e.g. Location, Taxon): Location
* Examples: `30` (reasonable lower limit **on or after 2020-05-01 of a GPS reading under good conditions if the actual accuracy estimate was not recorded at the time). `100` (reasonable lower limit before 2020-05-01 of a GPS reading under good conditions if the actual accuracy estimate was not recorded at the time).** `71` (uncertainty for a UTM coordinate having 100 meter precision and a known spatial reference system).
Submitted via https://github.com/gbif/doc-georeferencing-quick-reference-guide/issues/7 | 1.0 | Change term - coordinateUncertaintyInMeters - ## Change term
* Submitter: https://github.com/RicardoOrtizG
* Justification (why is this change necessary?): For completeness
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/terms/#dwc:coordinateUncertaintynMeters
Proposed new attributes of the term:
* Term name (in lowerCamelCase): coordinateUncertaintyInMeters
* Organized in Class (e.g. Location, Taxon): Location
* Examples: `30` (reasonable lower limit **on or after 2020-05-01 of a GPS reading under good conditions if the actual accuracy estimate was not recorded at the time). `100` (reasonable lower limit before 2020-05-01 of a GPS reading under good conditions if the actual accuracy estimate was not recorded at the time).** `71` (uncertainty for a UTM coordinate having 100 meter precision and a known spatial reference system).
Submitted via https://github.com/gbif/doc-georeferencing-quick-reference-guide/issues/7 | process | change term coordinateuncertaintyinmeters change term submitter justification why is this change necessary for completeness proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase coordinateuncertaintyinmeters organized in class e g location taxon location examples reasonable lower limit on or after of a gps reading under good conditions if the actual accuracy estimate was not recorded at the time reasonable lower limit before of a gps reading under good conditions if the actual accuracy estimate was not recorded at the time uncertainty for a utm coordinate having meter precision and a known spatial reference system submitted via | 1 |
14,198 | 17,099,225,637 | IssuesEvent | 2021-07-09 08:51:48 | prisma/prisma | https://api.github.com/repos/prisma/prisma | closed | Error: Error in migration engine. Reason: [libs/sql-schema-describer/src/walkers.rs:214:27] index out of bounds: the len is 0 but the index is 0 | bug/1-repro-available kind/bug process/candidate team/migrations | <!-- If required, please update the title to be clear and descriptive -->
Command: `prisma migrate save --experimental`
Version: `2.10.0`
Binary Version: `af1ae11a423edfb5d24092a85be11fa77c5e499c`
Report: https://prisma-errors.netlify.app/report/13412
OS: `x64 darwin 20.5.0`
JS Stacktrace:
```
Error: Error in migration engine.
Reason: [libs/sql-schema-describer/src/walkers.rs:214:27] index out of bounds: the len is 0 but the index is 0
Please create an issue in the migrate repo with
your `schema.prisma` and the prisma command you tried to use 🙏:
https://github.com/prisma/migrate/issues/new
at ChildProcess.<anonymous> (/Users/daeun/develop/payking-backend-service/node_modules/@prisma/cli/build/index.js:418:153)
at ChildProcess.emit (events.js:315:20)
at ChildProcess.EventEmitter.emit (domain.js:485:12)
at Process.ChildProcess._handle.onexit (internal/child_process.js:275:12)
```
Rust Stacktrace:
```
[libs/sql-schema-describer/src/walkers.rs:214:27] index out of bounds: the len is 0 but the index is 0
```
| 1.0 | Error: Error in migration engine. Reason: [libs/sql-schema-describer/src/walkers.rs:214:27] index out of bounds: the len is 0 but the index is 0 - <!-- If required, please update the title to be clear and descriptive -->
Command: `prisma migrate save --experimental`
Version: `2.10.0`
Binary Version: `af1ae11a423edfb5d24092a85be11fa77c5e499c`
Report: https://prisma-errors.netlify.app/report/13412
OS: `x64 darwin 20.5.0`
JS Stacktrace:
```
Error: Error in migration engine.
Reason: [libs/sql-schema-describer/src/walkers.rs:214:27] index out of bounds: the len is 0 but the index is 0
Please create an issue in the migrate repo with
your `schema.prisma` and the prisma command you tried to use 🙏:
https://github.com/prisma/migrate/issues/new
at ChildProcess.<anonymous> (/Users/daeun/develop/payking-backend-service/node_modules/@prisma/cli/build/index.js:418:153)
at ChildProcess.emit (events.js:315:20)
at ChildProcess.EventEmitter.emit (domain.js:485:12)
at Process.ChildProcess._handle.onexit (internal/child_process.js:275:12)
```
Rust Stacktrace:
```
[libs/sql-schema-describer/src/walkers.rs:214:27] index out of bounds: the len is 0 but the index is 0
```
| process | error error in migration engine reason index out of bounds the len is but the index is command prisma migrate save experimental version binary version report os darwin js stacktrace error error in migration engine reason index out of bounds the len is but the index is please create an issue in the migrate repo with your schema prisma and the prisma command you tried to use 🙏 at childprocess users daeun develop payking backend service node modules prisma cli build index js at childprocess emit events js at childprocess eventemitter emit domain js at process childprocess handle onexit internal child process js rust stacktrace index out of bounds the len is but the index is | 1 |
7,162 | 10,310,107,102 | IssuesEvent | 2019-08-29 14:33:34 | Graylog2/graylog2-server | https://api.github.com/repos/Graylog2/graylog2-server | closed | Pipeline Rule Editor: "Save and Close" | improvement processing triaged | When editing a pipeline rule, I often want to validate that it worked. For that I either use a search result or the simulator, both in a new tab.
Unfortunately, the only way to save a rule is by also being redirected back to the rule list. In many cases, I want to make more edits or fix what I did before but currently I have to find the rule again and open the rule editor.
It would be great to either have the option to choose between "Save and Close" and "Save and Keep Editing" or to simply make keeping the editor open the default. | 1.0 | Pipeline Rule Editor: "Save and Close" - When editing a pipeline rule, I often want to validate that it worked. For that I either use a search result or the simulator, both in a new tab.
Unfortunately, the only way to save a rule is by also being redirected back to the rule list. In many cases, I want to make more edits or fix what I did before but currently I have to find the rule again and open the rule editor.
It would be great to either have the option to choose between "Save and Close" and "Save and Keep Editing" or to simply make keeping the editor open the default. | process | pipeline rule editor save and close when editing a pipeline rule i often want to validate that it worked for that i either use a search result or the simulator both in a new tab unfortunately the only way to save a rule is by also being redirected back to the rule list in many cases i want to make more edits or fix what i did before but currently i have to find the rule again and open the rule editor it would be great to either have the option to choose between save and close and save and keep editing or to simply make keeping the editor open the default | 1 |
7,725 | 10,832,944,443 | IssuesEvent | 2019-11-11 11:47:05 | prisma/photonjs | https://api.github.com/repos/prisma/photonjs | closed | The findOne throws RecordDoesNotExist in alpha.282 | process/candidate | It should not throw such an error since preview v015. BTW in v015 it was working fine. | 1.0 | The findOne throws RecordDoesNotExist in alpha.282 - It should not throw such an error since preview v015. BTW in v015 it was working fine. | process | the findone throws recorddoesnotexist in alpha it should not throw such an error since preview btw in it was working fine | 1 |
13,397 | 2,755,409,489 | IssuesEvent | 2015-04-26 16:33:57 | benklett/gwen-dotnet | https://api.github.com/repos/benklett/gwen-dotnet | closed | Unable to Run Sample SFML | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Download a fresh copy of the Gwen dotNet project
2. Copy the csfml dlls
3. Run the solution
What is the expected output? What do you see instead?
I expect to see the Unit test window with the controls
All i get is an error:
"An unhandled exception of type 'System.BadImageFormatException' occurred in
Unknown Module.
Additional information: Could not load file or assembly 'sfmlnet-graphics-2,
Version=0.0.0.0, Culture=neutral, PublicKeyToken=null' or one of its
dependencies. An attempt was made to load a program with an incorrect format."
What version of the product are you using? On what operating system?
Using VS Community 2013
Windows 10
```
Original issue reported on code.google.com by `Scello...@gmail.com` on 22 Dec 2014 at 12:14 | 1.0 | Unable to Run Sample SFML - ```
What steps will reproduce the problem?
1. Download a fresh copy of the Gwen dotNet project
2. Copy the csfml dlls
3. Run the solution
What is the expected output? What do you see instead?
I expect to see the Unit test window with the controls
All i get is an error:
"An unhandled exception of type 'System.BadImageFormatException' occurred in
Unknown Module.
Additional information: Could not load file or assembly 'sfmlnet-graphics-2,
Version=0.0.0.0, Culture=neutral, PublicKeyToken=null' or one of its
dependencies. An attempt was made to load a program with an incorrect format."
What version of the product are you using? On what operating system?
Using VS Community 2013
Windows 10
```
Original issue reported on code.google.com by `Scello...@gmail.com` on 22 Dec 2014 at 12:14 | non_process | unable to run sample sfml what steps will reproduce the problem download a fresh copy of the gwen dotnet project copy the csfml dlls run the solution what is the expected output what do you see instead i expect to see the unit test window with the controls all i get is an error an unhandled exception of type system badimageformatexception occurred in unknown module additional information could not load file or assembly sfmlnet graphics version culture neutral publickeytoken null or one of its dependencies an attempt was made to load a program with an incorrect format what version of the product are you using on what operating system using vs community windows original issue reported on code google com by scello gmail com on dec at | 0 |
5,376 | 8,203,956,374 | IssuesEvent | 2018-09-03 03:03:41 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | pipeline-ui loading failed over https | area/pipeline kind/bug process/needs-ui team/cn version/1.6 | ## URL
` https://<domain>/r/projects/1a5/pipeline-ui/#/fail`
## Console Error
`Mixed Content: The page at 'https://<domain>/r/projects/1a5/pipeline-ui/' was loaded over HTTPS, but requested an insecure XMLHttpRequest endpoint 'http://<domain>/r/projects/1a5/pipeline-server:60080/v1/pipelines?limit=-1&sort=name'. This request has been blocked; the content must be served over HTTPS.` | 1.0 | pipeline-ui loading failed over https - ## URL
` https://<domain>/r/projects/1a5/pipeline-ui/#/fail`
## Console Error
`Mixed Content: The page at 'https://<domain>/r/projects/1a5/pipeline-ui/' was loaded over HTTPS, but requested an insecure XMLHttpRequest endpoint 'http://<domain>/r/projects/1a5/pipeline-server:60080/v1/pipelines?limit=-1&sort=name'. This request has been blocked; the content must be served over HTTPS.` | process | pipeline ui loading failed over https url console error mixed content the page at was loaded over https but requested an insecure xmlhttprequest endpoint this request has been blocked the content must be served over https | 1 |
225,715 | 17,287,312,547 | IssuesEvent | 2021-07-24 01:27:07 | anagainaru/HPC_IOpatterns | https://api.github.com/repos/anagainaru/HPC_IOpatterns | opened | Darshan alternatives | documentation | Recorder 2.0 ([https://snir.cs.illinois.edu/listed/Recorder_2_0.pdf](https://snir.cs.illinois.edu/listed/Recorder_2_0.pdf))
TAU IO tracer
<img width="915" alt="Screen Shot 2021-07-23 at 9 25 46 PM" src="https://user-images.githubusercontent.com/16229479/126853649-894d192c-9ef7-4027-bae0-afb113e49ecc.png">
Add metadata information and separation on files | 1.0 | Darshan alternatives - Recorder 2.0 ([https://snir.cs.illinois.edu/listed/Recorder_2_0.pdf](https://snir.cs.illinois.edu/listed/Recorder_2_0.pdf))
TAU IO tracer
<img width="915" alt="Screen Shot 2021-07-23 at 9 25 46 PM" src="https://user-images.githubusercontent.com/16229479/126853649-894d192c-9ef7-4027-bae0-afb113e49ecc.png">
Add metadata information and separation on files | non_process | darshan alternatives recorder tau io tracer img width alt screen shot at pm src add metadata information and separation on files | 0 |
415,002 | 12,122,791,432 | IssuesEvent | 2020-04-22 11:36:47 | onaio/reveal-frontend | https://api.github.com/repos/onaio/reveal-frontend | closed | The Conditional formatting rules percentage table shows double | Priority: High bug | Login to the system
Select the IRS reporting
Select one of the IRS plans
Observe the Conditional formatting rules percentage table.
Observe if it shows double data. It should not show two sets of data.

| 1.0 | The Conditional formatting rules percentage table shows double - Login to the system
Select the IRS reporting
Select one of the IRS plans
Observe the Conditional formatting rules percentage table.
Observe if it shows double data. It should not show two sets of data.

| non_process | the conditional formatting rules percentage table shows double login to the system select the irs reporting select one of the irs plans observe the conditional formatting rules percentage table observe if it shows double data it should not show two sets of data | 0 |
19,837 | 26,234,907,040 | IssuesEvent | 2023-01-05 05:58:34 | hsmusic/hsmusic-data | https://api.github.com/repos/hsmusic/hsmusic-data | opened | Wallpapers for the 'classic' Homestuck volumes | scope: official type: addition type: involved process what: art & layout media | Homestuck Vol. 4 (as well as its rerelease Vol. 1-4) released with an exclusive SBURB-themed wallpaper. We added it to the wiki with the additional files feature, but it never felt 100% appropriate to have a background only for Vol. 4 and not the other three initial albums (although it looks really good as an album background). I'd like to use some of [MSPA's wallpapers section](http://www.mspaintadventures.com/desktops.html) for this since they match the SBURB wallpaper's early Homestuck feel. The idea is for each volume to have a wallpaper of the appropriate kid's house and for Vol. 1-4 to use the SBURB wallpaper.
- [ ] Vol. 1
- [ ] Trawl through flashes for appropriate HQ assets
- [ ] Vol. 2
- [ ] Trawl through flashes for appropriate HQ assets
- [x] [Vol. 3](http://www.mspaintadventures.com/desktops/daveapt_1920x1080.jpg)
- [ ] [Vol. 4](http://www.mspaintadventures.com/desktops/jadehouse_1920x1080.jpg)
- [ ] Carefully fill in blank edges
- [ ] [Vol. 1-4](https://hsmusic.wiki/media/album-additional/homestuck-vol-4/sburbwp_1920x1080.jpg)
- [ ] Slightly adjust horizontally to prevent scrollbar from mucking with symmetry (5px to the left)
- [ ] Actually there's probably a way to avoid this in CSS @towerofnix help? | 1.0 | Wallpapers for the 'classic' Homestuck volumes - Homestuck Vol. 4 (as well as its rerelease Vol. 1-4) released with an exclusive SBURB-themed wallpaper. We added it to the wiki with the additional files feature, but it never felt 100% appropriate to have a background only for Vol. 4 and not the other three initial albums (although it looks really good as an album background). I'd like to use some of [MSPA's wallpapers section](http://www.mspaintadventures.com/desktops.html) for this since they match the SBURB wallpaper's early Homestuck feel. The idea is for each volume to have a wallpaper of the appropriate kid's house and for Vol. 1-4 to use the SBURB wallpaper.
- [ ] Vol. 1
- [ ] Trawl through flashes for appropriate HQ assets
- [ ] Vol. 2
- [ ] Trawl through flashes for appropriate HQ assets
- [x] [Vol. 3](http://www.mspaintadventures.com/desktops/daveapt_1920x1080.jpg)
- [ ] [Vol. 4](http://www.mspaintadventures.com/desktops/jadehouse_1920x1080.jpg)
- [ ] Carefully fill in blank edges
- [ ] [Vol. 1-4](https://hsmusic.wiki/media/album-additional/homestuck-vol-4/sburbwp_1920x1080.jpg)
- [ ] Slightly adjust horizontally to prevent scrollbar from mucking with symmetry (5px to the left)
- [ ] Actually there's probably a way to avoid this in CSS @towerofnix help? | process | wallpapers for the classic homestuck volumes homestuck vol as well as its rerelease vol released with an exclusive sburb themed wallpaper we added it to the wiki with the additional files feature but it never felt appropriate to have a background only for vol and not the other three initial albums although it looks really good as an album background i d like to use some of for this since they match the sburb wallpaper s early homestuck feel the idea is for each volume to have a wallpaper of the appropriate kid s house and for vol to use the sburb wallpaper vol trawl through flashes for appropriate hq assets vol trawl through flashes for appropriate hq assets carefully fill in blank edges slightly adjust horizontally to prevent scrollbar from mucking with symmetry to the left actually there s probably a way to avoid this in css towerofnix help | 1 |
218,151 | 24,351,797,236 | IssuesEvent | 2022-10-03 01:20:31 | hygieia/hygieia-relateditems-collector | https://api.github.com/repos/hygieia/hygieia-relateditems-collector | opened | CVE-2022-42004 (Medium) detected in jackson-databind-2.5.0.jar | security vulnerability | ## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.0/jackson-databind-2.5.0.jar</p>
<p>
Dependency Hierarchy:
- core-3.7.11.jar (Root Library)
- spring-boot-starter-web-1.3.0.RELEASE.jar
- :x: **jackson-databind-2.5.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hygieia/hygieia-relateditems-collector/commit/4c8d5a1732722418ec10d2f1f5fd7e7a6e43c83b">4c8d5a1732722418ec10d2f1f5fd7e7a6e43c83b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-42004 (Medium) detected in jackson-databind-2.5.0.jar - ## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.0/jackson-databind-2.5.0.jar</p>
<p>
Dependency Hierarchy:
- core-3.7.11.jar (Root Library)
- spring-boot-starter-web-1.3.0.RELEASE.jar
- :x: **jackson-databind-2.5.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hygieia/hygieia-relateditems-collector/commit/4c8d5a1732722418ec10d2f1f5fd7e7a6e43c83b">4c8d5a1732722418ec10d2f1f5fd7e7a6e43c83b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy core jar root library spring boot starter web release jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in beandeserializer deserializefromarray to prevent use of deeply nested arrays an application is vulnerable only with certain customized choices for deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with mend | 0 |
73,093 | 15,252,478,731 | IssuesEvent | 2021-02-20 03:05:41 | AlexRogalskiy/github-action-quotes | https://api.github.com/repos/AlexRogalskiy/github-action-quotes | opened | CVE-2020-28500 (Medium) detected in lodash-4.17.20.tgz | security vulnerability | ## CVE-2020-28500 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p>
<p>Path to dependency file: github-action-quotes/package.json</p>
<p>Path to vulnerable library: github-action-quotes/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commit-analyzer-6.3.3.tgz (Root Library)
- :x: **lodash-4.17.20.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-quotes/commit/0baf056b30e1d67519c35e91b907bfe181ecba6d">0baf056b30e1d67519c35e91b907bfe181ecba6d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Steps to reproduce (provided by reporter Liyuan Chen): var lo = require('lodash'); function build_blank (n) { var ret = "1" for (var i = 0; i < n; i++) { ret += " " } return ret + "1"; } var s = build_blank(50000) var time0 = Date.now(); lo.trim(s) var time_cost0 = Date.now() - time0; console.log("time_cost0: " + time_cost0) var time1 = Date.now(); lo.toNumber(s) var time_cost1 = Date.now() - time1; console.log("time_cost1: " + time_cost1) var time2 = Date.now(); lo.trimEnd(s) var time_cost2 = Date.now() - time2; console.log("time_cost2: " + time_cost2)
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-28500 (Medium) detected in lodash-4.17.20.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p>
<p>Path to dependency file: github-action-quotes/package.json</p>
<p>Path to vulnerable library: github-action-quotes/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commit-analyzer-6.3.3.tgz (Root Library)
- :x: **lodash-4.17.20.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-quotes/commit/0baf056b30e1d67519c35e91b907bfe181ecba6d">0baf056b30e1d67519c35e91b907bfe181ecba6d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Steps to reproduce (provided by reporter Liyuan Chen): var lo = require('lodash'); function build_blank (n) { var ret = "1" for (var i = 0; i < n; i++) { ret += " " } return ret + "1"; } var s = build_blank(50000) var time0 = Date.now(); lo.trim(s) var time_cost0 = Date.now() - time0; console.log("time_cost0: " + time_cost0) var time1 = Date.now(); lo.toNumber(s) var time_cost1 = Date.now() - time1; console.log("time_cost1: " + time_cost1) var time2 = Date.now(); lo.trimEnd(s) var time_cost2 = Date.now() - time2; console.log("time_cost2: " + time_cost2)
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file github action quotes package json path to vulnerable library github action quotes node modules lodash package json dependency hierarchy commit analyzer tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions steps to reproduce provided by reporter liyuan chen var lo require lodash function build blank n var ret for var i i n i ret return ret var s build blank var date now lo trim s var time date now console log time time var date now lo tonumber s var time date now console log time time var date now lo trimend s var time date now console log time time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href step up your open source security game with whitesource | 0 |
17 | 2,496,244,244 | IssuesEvent | 2015-01-06 18:05:45 | vivo-isf/vivo-isf-ontology | https://api.github.com/repos/vivo-isf/vivo-isf-ontology | closed | potassium ion transport | biological_process imported | _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on December 12, 2012 15:44:59_
GO:0006813
parent: transport
\<a href="http://purl.obolibrary.org/obo/GO_0006810" rel="nofollow">http://purl.obolibrary.org/obo/GO_0006810</a>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=174_ | 1.0 | potassium ion transport - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on December 12, 2012 15:44:59_
GO:0006813
parent: transport
\<a href="http://purl.obolibrary.org/obo/GO_0006810" rel="nofollow">http://purl.obolibrary.org/obo/GO_0006810</a>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=174_ | process | potassium ion transport from on december go parent transport original issue | 1 |
238,501 | 7,780,053,555 | IssuesEvent | 2018-06-05 18:46:56 | mobileappdevhm/team2 | https://api.github.com/repos/mobileappdevhm/team2 | closed | Server with REST API with web management interface | high priority | We need a server with rest API and database so that we can access **all** resources: Courses, Favorites, etc.
Additionally we need a web interface to manage courses, etc and the selection result. | 1.0 | Server with REST API with web management interface - We need a server with rest API and database so that we can access **all** resources: Courses, Favorites, etc.
Additionally we need a web interface to manage courses, etc and the selection result. | non_process | server with rest api with web management interface we need a server with rest api and database so that we can access all resources courses favorites etc additionally we need a web interface to manage courses etc and the selection result | 0 |
93,677 | 15,895,518,140 | IssuesEvent | 2021-04-11 14:16:44 | rammatzkvosky/123 | https://api.github.com/repos/rammatzkvosky/123 | opened | CVE-2020-14195 (High) detected in jackson-databind-2.8.8.jar | security vulnerability | ## CVE-2020-14195 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: 123/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/123/commit/d30fca67e915548a9a9d8fbcfe3506cf080ce3b5">d30fca67e915548a9a9d8fbcfe3506cf080ce3b5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).
<p>Publish Date: 2020-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195>CVE-2020-14195</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195</a></p>
<p>Release Date: 2020-06-16</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.5</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.5"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-14195","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-14195 (High) detected in jackson-databind-2.8.8.jar - ## CVE-2020-14195 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: 123/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/123/commit/d30fca67e915548a9a9d8fbcfe3506cf080ce3b5">d30fca67e915548a9a9d8fbcfe3506cf080ce3b5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).
<p>Publish Date: 2020-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195>CVE-2020-14195</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195</a></p>
<p>Release Date: 2020-06-16</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.5</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.5"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-14195","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_process | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org jsecurity realm jndi jndirealmfactory aka org jsecurity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org jsecurity realm jndi jndirealmfactory aka org jsecurity vulnerabilityurl | 0 |
3,674 | 6,706,650,443 | IssuesEvent | 2017-10-12 08:01:16 | openvstorage/framework | https://api.github.com/repos/openvstorage/framework | closed | Make it possible to add more than 1 DB disk | process_wontfix type_enhancement | If its possible to add more than 1 DB disk you will be able to spread your arakoons across those disks. When a DB disk would fail, only a part of the arakoons fail. | 1.0 | Make it possible to add more than 1 DB disk - If its possible to add more than 1 DB disk you will be able to spread your arakoons across those disks. When a DB disk would fail, only a part of the arakoons fail. | process | make it possible to add more than db disk if its possible to add more than db disk you will be able to spread your arakoons across those disks when a db disk would fail only a part of the arakoons fail | 1 |
102,706 | 12,821,374,356 | IssuesEvent | 2020-07-06 07:57:33 | greatnewcls/KNLWKKGOTE62XGFSVAZIMVUH | https://api.github.com/repos/greatnewcls/KNLWKKGOTE62XGFSVAZIMVUH | reopened | iiMnIRzmoirW6ZpXMFbXJCl+thoR3RNe94CXMplFXK1p4CfJhUYjftKNyLO+RqWa51fGGJoq3yOMqTjLkNe+lMb2VtRFR/SVFeuVZrzJew8cjCMkfIHasUOJD+zHdrhl3yQQm9m7lWcj/TA6Axzb97yDF2D4sNYwoX4iHHtkvQc= | design | SmUQK7lBu+g2PkMB73ijExIIwfUMC8alTFNocU6K+akOsR3p6mRXaH0kZ6YTWC+rH9PEHBIGgDmEmYEBqWDp+GRFsTb/U1qGWSksus5RlBSiEdjJ8yHeANCI21JC0gwpEgjB9QwLxqVMU2hxTor5qT9c0HJ8UlyMHBX/MDXksLcf08QcEgaAOYSZgQGpYOn4HHyBoKn/hhI62y4mHbAyhjj9bE5jmRMCCkNIAMSxT7USCMH1DAvGpUxTaHFOivmp447CxTtIAY7NWPkd8YfeFh/TxBwSBoA5hJmBAalg6fgmQLMuuCv0/mt83SDHTXvaJaeL1xBu2jzuBkxTS5pV2T79kw9n4yaxGWhTR/bfgm2gm6Fgp0Os5FTDMRksst/0H9PEHBIGgDmEmYEBqWDp+Js1fXfoxgiZVpJVGuljgpsOXHFWAR0ezOldHh09Hc5phlhfsVu1/jgZA0L5c7PsWikHw7QOihN0nQnP9acCUXEf08QcEgaAOYSZgQGpYOn4BtRLCcDbcV4YvQV2oTlcs00hOjXxZKamIxk6j9uyiFCbIExjxqvAbyy5E+i81K0EEN9T5enpnKTsnqSOhcUPwx/TxBwSBoA5hJmBAalg6fhc2gqglIdV3O1T+f/VIDDTfCa5yq6bcFd1DVzD70LAnfRjoT2cd5nIXHGBuunKpxDuNCkBFrMgpcSHSebhm8kJH9PEHBIGgDmEmYEBqWDp+NxL4CWDgow5pRR34frOumCiXT6TcJL39XBnlVEvdTD8Xqx0cYukLJlgA6Q43oi3W5iUQVsd60fkTQUbkU5R5fMf08QcEgaAOYSZgQGpYOn42MEvpHK7h5mTJDcChdBhBryQ7oLtEBOlX52b4Dc3bGASCMH1DAvGpUxTaHFOivmpUKVMLQWV4fncGAi9CI6n+x/TxBwSBoA5hJmBAalg6fhS035hreywFL1UswooupwB8tsQuZIV97GNcCihH6Rj9hIIwfUMC8alTFNocU6K+amZ1uL5mqVpP5aDx/FnklLhH9PEHBIGgDmEmYEBqWDp+BP13iLYs1jJGi4u5yelr9BjhlmQe0LqK+lOyYtoGDbpEgjB9QwLxqVMU2hxTor5qQ3hgVMg1kRi91TzH8z11Msf08QcEgaAOYSZgQGpYOn4g4u0676fm4GcDXRDlGyLlSvMB1D2HqStnndX9k0o9Q4jb3PM6MxZAwEtKiVRelYaKMBnGyh9c27HXckoh/3CEB/TxBwSBoA5hJmBAalg6fheY+SWZp9giu3VZn289QCcJU3s0vI7srBXfhGTm9PazBIIwfUMC8alTFNocU6K+akUqOcKoinbKeUJh/PTov79H9PEHBIGgDmEmYEBqWDp+G4bStbJdDEdtDhqqb7YePGfTpK8PL/gKDOUM1KGiWWfhlhfsVu1/jgZA0L5c7PsWkO1NySUMNFh6fnkmuYUY0Yf08QcEgaAOYSZgQGpYOn4Bjc1g6gczeaZCDUbON0o94vN3K6LjVvDel0hxNNl2hUSCMH1DAvGpUxTaHFOivmp1xBZ62JCfBj7hhF29APmNh/TxBwSBoA5hJmBAalg6fhJjSOt0VEq2OCOO8aNi5FrJyS/OlcQ53Ow0W/fX9lPehIIwfUMC8alTFNocU6K+amPN8p5QZcvJsYBAtTESMlJH9PEHBIGgDmEmYEBqWDp+E/3Ju6O5fnm+JovHmoe10fkMC8xkYoRBnEd/VVeSjnhcKuyI4lPljksjRaQxz7DZVu/agf/y5MdcpqbYNE+u+kf08QcEgaAOYSZgQGpYOn4Ubd2+0bnC3AmH3wCpr9c+6IR2MnzId4A0IjbUkLSDCkSCMH1DAvGpUxTaHFOivmp3FCtlALoErAkoW3E+CMjBx/TxBwSBoA5hJmBAalg6fiWDonqkox8Hitk6UwNQH/ZOP1sTmOZEwIKQ0gAxLFPtRIIwfUMC8alTFNocU6K+an53ctiWXBCjuXZbYQY5YrLH9PEHBIGgDmEmYEBqWDp+DsNwW5i5v4+297UTUGk6aOkMq/uCOtHXJMWzXd582YahIk3qQjg0LgSOFGqzzOQKxv3WzYJrLaPLmkW0vtWS2Ef08QcEgaAOYSZgQGpYOn4Mu6gcRmP5W8pjDh8eTewwSckvzpXEOdzsNFv31/ZT3oSCMH1DAvGpUxTaHFOivmpWyrzt71xrONeeCOw+8T3mx/TxBwSBoA5hJmBAalg6fhvJ6NsnR3q4+XE9/F2KRKYJyS/OlcQ53Ow0W/fX9lPehIIwfUMC8alTFNocU6K+alcYquz1bLdJQuqbN8QlWgxH9PEHBIGgDmEmYEBqWDp+Eo6ukSDqgjjuZDKOWdYnFc4/WxOY5kTAgpDSADEsU+1EgjB9QwLxqVMU2hxTor5qYWHlj8hZYg4nDaH49HG3jQf08QcEgaAOYSZgQGpYOn4RyIRCv52pzsA9pKTbyFnuqr/ZeuBwRqNM/4PiodkEmO7OraWB16p6cJ1TEngcuXiFxvgAhp2CUjDtZIICn1Rxh/TxBwSBoA5hJmBAalg6fiTuhcC8pOMiuDs8bHYqEMmVe84UlBVIY7XZEWZ8XJBkxBpixGjXI1kf5luol5+eEZVI8K69XXozwMCBhCnBL/wH9PEHBIGgDmEmYEBqWDp+OvADpBAP6GNYQ8k33g5yS+iEdjJ8yHeANCI21JC0gwpEgjB9QwLxqVMU2hxTor5qUuMo68Gi7KWmbERE+eJaDYf08QcEgaAOYSZgQGpYOn4mKvSHnf32T8ov19P3Eyk2CckvzpXEOdzsNFv31/ZT3oSCMH1DAvGpUxTaHFOivmp44wUl7juS67tubmIHL5Hoh/TxBwSBoA5hJmBAalg6fj9JG0MFJmE3ZXNjMFy0CHZvq3g7+J4//u4z7Ab50QAI1d2Zt534RgWQyZUmDRQTkE1MtAkD5JllHam501uxiRsH9PEHBIGgDmEmYEBqWDp+L51yVaTijwyoobVjIjFXGXy2xC5khX3sY1wKKEfpGP2EgjB9QwLxqVMU2hxTor5qW40ibNYIvfSlrXWmbnaEdwf08QcEgaAOYSZgQGpYOn4nNYmwNZg1UMMLaWA/QP+gL6t4O/ieP/7uM+wG+dEACNXdmbed+EYFkMmVJg0UE5BrknhvT+q5Yhl9W+pzIZ4sx/TxBwSBoA5hJmBAalg6fjRVI9fuLLXvyOoPeR5mNiV8tsQuZIV97GNcCihH6Rj9hIIwfUMC8alTFNocU6K+anwSbnn3kbpxJtNE8uGhugGH9PEHBIGgDmEmYEBqWDp+PpF/cjo3vVkhKWqK2RSkuXTfpAXHkI010RLRiOR4MlYV3Zm3nfhGBZDJlSYNFBOQW+uCwgPpRYpCCXuBonhKOwf08QcEgaAOYSZgQGpYOn4U/Z8OcWqibRp88nIdhKbU2OGWZB7Quor6U7Ji2gYNukSCMH1DAvGpUxTaHFOivmpcLihLDS4O2o9hoc007WLQx/TxBwSBoA5hJmBAalg6fgSV98BayNefL7mlT4TTXe1aA4GPmSfXzsJ7dYEFhA4DPRjoT2cd5nIXHGBuunKpxATH+Nn5ULeoleHTFADwi65H9PEHBIGgDmEmYEBqWDp+JyWNLA2poaF98erzR8Cetc0Uuwm0uik5Ue+2ZZnc6AyVjH/LpFtIJkEm5KuK+k4+AxSQO3qFAroekFv2KRJlw0f08QcEgaAOYSZgQGpYOn4lyGfgvidpuuCygilRY2fPMdxo3Jqw9hnq96WL5pRRZcSCMH1DAvGpUxTaHFOivmpgHC48E44LxBqc4K0rdOyOx/TxBwSBoA5hJmBAalg6fjyBdPQYzrnIenb5bfX6tYX8tsQuZIV97GNcCihH6Rj9hIIwfUMC8alTFNocU6K+akZ7kBQ1NwsCZbDvVxcgih5H9PEHBIGgDmEmYEBqWDp+Acf4cjuqbW8SyHrQd0OVb+iEdjJ8yHeANCI21JC0gwpEgjB9QwLxqVMU2hxTor5qfc/alJsE2joWaRMDIUNZkkf08QcEgaAOYSZgQGpYOn4bAhhN1QeK+ywrvRoL5lk0CckvzpXEOdzsNFv31/ZT3oSCMH1DAvGpUxTaHFOivmpN1ejVkIMmCea+MgnS5Xnnx/TxBwSBoA5hJmBAalg6fjASKy33n3gxNg9wOfk3x0lY4ZZkHtC6ivpTsmLaBg26RIIwfUMC8alTFNocU6K+anPqc55WZIcTAKHsFU/Av9PH9PEHBIGgDmEmYEBqWDp+DYIj6tgzjOKODoAP26ZzSn4V0hOZAvicWWSnHKxYiHpV3Zm3nfhGBZDJlSYNFBOQf6/dQr/8TzVMNfvU8buGWgf08QcEgaAOYSZgQGpYOn4Mw29DL7kasPmvjSGa5chQ2OGWZB7Quor6U7Ji2gYNukSCMH1DAvGpUxTaHFOivmpEVVTzGtSACted8UBKa6Fux/TxBwSBoA5hJmBAalg6fiuI2bI7tHBWvHwhi6jtWSk2r13erpF+dSnkMkV7/PguxIIwfUMC8alTFNocU6K+alBsXdeCeEy88KLiIl0uQtLH9PEHBIGgDmEmYEBqWDp+GfRP+5uwgtUDF0+d9xqiKf4V0hOZAvicWWSnHKxYiHpV3Zm3nfhGBZDJlSYNFBOQQ5t+CaHqPtwKiHNV8YZOPkf08QcEgaAOYSZgQGpYOn4k9+AQek1xVK9mjfRJ1HmC+U8R31LKcQ/w4AFLN/JJTASCMH1DAvGpUxTaHFOivmps7rke53BsWqPqGoENIerbR/TxBwSBoA5hJmBAalg6fja4mj24zucDMCHE3bZ8X0gohHYyfMh3gDQiNtSQtIMKRIIwfUMC8alTFNocU6K+akB2jtbs0Pnd8zlKjQXC7MyH9PEHBIGgDmEmYEBqWDp+IbLnrjadFCbb+qW8YHAyx5H3wqv3wyq5Yrl8PbUij+3V3Zm3nfhGBZDJlSYNFBOQSedUx43MhSp8p0Ts+E0vwkf08QcEgaAOYSZgQGpYOn4QC41mzDUae7PGOc1UqguA0ffCq/fDKrliuXw9tSKP7dXdmbed+EYFkMmVJg0UE5BXr3Xzsjb12fOZOJum5ZcyR/TxBwSBoA5hJmBAalg6fhn5LBPmksGyKs3ue1/Wy+gxZFiuWgqKMTwlSVc+w+bn4KYayF+X3onpdWm8JzGxjvBp+9kU7BtRwacfGguk7BgH9PEHBIGgDmEmYEBqWDp+EciEQr+dqc7APaSk28hZ7outhmmLQVrPblwXnxVTFaW17SvoyZpa6+fyZv93e88ypgkMhZvCzwmTMUPGN3nRRAf08QcEgaAOYSZgQGpYOn4xVMim6/A0SpXr703tOg5g4xXqZjrgiLycXggrLvK0GuCmGshfl96J6XVpvCcxsY7cHSWZL5zuqj8K7NVDZOUvx/TxBwSBoA5hJmBAalg6fg+0bN/WLrhKWqemDnziNJcaD6WgwomxjRrZFV32wZy2qhXlgrpD7TLqHA2z09HkaIaudq69bdXxT5ZndK09DatH9PEHBIGgDmEmYEBqWDp+BYQAcUF+13XAUfGXZyiZDQ+++7JTdGH5Zmvf8OJr6wrAxBCQqQ6GD0PJYrp5eg501yO/HQX7KbMbfUx9o6bmzkf08QcEgaAOYSZgQGpYOn4ZAIy2nycs3ymZVVmYMI566IR2MnzId4A0IjbUkLSDCkSCMH1DAvGpUxTaHFOivmp82xSnwuf7egZlzRJ8oVOJx/TxBwSBoA5hJmBAalg6fiN/vK6lKKPhwoRG9fbnFzCrkmvul111U1SNzwlX+9qpK0rwiIYWzsCTct59Ky9aQfC0ooOlMjjbntCx9N+MLNnH9PEHBIGgDmEmYEBqWDp+OZ/YmP3iodn7u/ezlWXKONdXVyl5XkBpuFL8ts6ciF6K3tiqSoJkLf4PBmccc7ECniHBqA7vwhm5nEHuTUQjFD9jQvdIaEvNeRqPXi6kUkiOPE9faWIl/hjy2eyGB1KVE+ETfaU9UTnO3rv6C15dN2bIExjxqvAbyy5E+i81K0E6rSybYkivFPsFMWyx2kagP2NC90hoS815Go9eLqRSSIj5Qgo8XER6KNGhK6mvVC6Y4ZZkHtC6ivpTsmLaBg26RIIwfUMC8alTFNocU6K+alK/sFynN6Vn2GQf2QASp8l/Y0L3SGhLzXkaj14upFJImOqra5JIlgcyt/5qzBX4x1jhlmQe0LqK+lOyYtoGDbpEgjB9QwLxqVMU2hxTor5qYRrS5Mi5ZJA8IWU4OjsRMP9jQvdIaEvNeRqPXi6kUki9GxVVI96sAaUZK+yMbipe5pq2X09R/Amhz+F67g6AxybIExjxqvAbyy5E+i81K0EYQjrmvqzXAhZrFk3pLdetv2NC90hoS815Go9eLqRSSKmRMx7TpyekyTFwPUmXjrnLn5fqwOrZ5nrOWW4sBCX9M1z3SQj90R4DLm90AAFmspNizwCE3V0X0QCGC7qrk57/Y0L3SGhLzXkaj14upFJIkngBLuxlZakahk6qsh0yvAadKYAARFyVvM7kQ6DoTER5J9/tyieL+qKg78Kj2vYTLpoQdreQ2uP+cfPpNfMZy/9jQvdIaEvNeRqPXi6kUki0AJo8edrJl1SkeZ4RQC7kYkfvnSQNQHM2WdhEowQ2O033FcLjTse3oPBOfQHcQ9rqiJYfhzyRX2klBRRvL8M5B/TxBwSBoA5hJmBAalg6fhXps5LDStEcd0cil/kLiU0JyS/OlcQ53Ow0W/fX9lPehIIwfUMC8alTFNocU6K+anGD8twxh7sDomA5Kz0PdgvH9PEHBIGgDmEmYEBqWDp+BRsSlub2Ul0S7c6FDAXno72EQ7mD6LRBdzgrBma7EDsmyBMY8arwG8suRPovNStBDyAc7k/g3AfTj+x7h/Shyz9jQvdIaEvNeRqPXi6kUki20VlAm4qY0su0eGKp3RCK2RuBD3ohuPg+c4bs7W8VaIqf+RfSsPtLq56+DVgXXo0 | 1.0 | iiMnIRzmoirW6ZpXMFbXJCl+thoR3RNe94CXMplFXK1p4CfJhUYjftKNyLO+RqWa51fGGJoq3yOMqTjLkNe+lMb2VtRFR/SVFeuVZrzJew8cjCMkfIHasUOJD+zHdrhl3yQQm9m7lWcj/TA6Axzb97yDF2D4sNYwoX4iHHtkvQc= - SmUQK7lBu+g2PkMB73ijExIIwfUMC8alTFNocU6K+akOsR3p6mRXaH0kZ6YTWC+rH9PEHBIGgDmEmYEBqWDp+GRFsTb/U1qGWSksus5RlBSiEdjJ8yHeANCI21JC0gwpEgjB9QwLxqVMU2hxTor5qT9c0HJ8UlyMHBX/MDXksLcf08QcEgaAOYSZgQGpYOn4HHyBoKn/hhI62y4mHbAyhjj9bE5jmRMCCkNIAMSxT7USCMH1DAvGpUxTaHFOivmp447CxTtIAY7NWPkd8YfeFh/TxBwSBoA5hJmBAalg6fgmQLMuuCv0/mt83SDHTXvaJaeL1xBu2jzuBkxTS5pV2T79kw9n4yaxGWhTR/bfgm2gm6Fgp0Os5FTDMRksst/0H9PEHBIGgDmEmYEBqWDp+Js1fXfoxgiZVpJVGuljgpsOXHFWAR0ezOldHh09Hc5phlhfsVu1/jgZA0L5c7PsWikHw7QOihN0nQnP9acCUXEf08QcEgaAOYSZgQGpYOn4BtRLCcDbcV4YvQV2oTlcs00hOjXxZKamIxk6j9uyiFCbIExjxqvAbyy5E+i81K0EEN9T5enpnKTsnqSOhcUPwx/TxBwSBoA5hJmBAalg6fhc2gqglIdV3O1T+f/VIDDTfCa5yq6bcFd1DVzD70LAnfRjoT2cd5nIXHGBuunKpxDuNCkBFrMgpcSHSebhm8kJH9PEHBIGgDmEmYEBqWDp+NxL4CWDgow5pRR34frOumCiXT6TcJL39XBnlVEvdTD8Xqx0cYukLJlgA6Q43oi3W5iUQVsd60fkTQUbkU5R5fMf08QcEgaAOYSZgQGpYOn42MEvpHK7h5mTJDcChdBhBryQ7oLtEBOlX52b4Dc3bGASCMH1DAvGpUxTaHFOivmpUKVMLQWV4fncGAi9CI6n+x/TxBwSBoA5hJmBAalg6fhS035hreywFL1UswooupwB8tsQuZIV97GNcCihH6Rj9hIIwfUMC8alTFNocU6K+amZ1uL5mqVpP5aDx/FnklLhH9PEHBIGgDmEmYEBqWDp+BP13iLYs1jJGi4u5yelr9BjhlmQe0LqK+lOyYtoGDbpEgjB9QwLxqVMU2hxTor5qQ3hgVMg1kRi91TzH8z11Msf08QcEgaAOYSZgQGpYOn4g4u0676fm4GcDXRDlGyLlSvMB1D2HqStnndX9k0o9Q4jb3PM6MxZAwEtKiVRelYaKMBnGyh9c27HXckoh/3CEB/TxBwSBoA5hJmBAalg6fheY+SWZp9giu3VZn289QCcJU3s0vI7srBXfhGTm9PazBIIwfUMC8alTFNocU6K+akUqOcKoinbKeUJh/PTov79H9PEHBIGgDmEmYEBqWDp+G4bStbJdDEdtDhqqb7YePGfTpK8PL/gKDOUM1KGiWWfhlhfsVu1/jgZA0L5c7PsWkO1NySUMNFh6fnkmuYUY0Yf08QcEgaAOYSZgQGpYOn4Bjc1g6gczeaZCDUbON0o94vN3K6LjVvDel0hxNNl2hUSCMH1DAvGpUxTaHFOivmp1xBZ62JCfBj7hhF29APmNh/TxBwSBoA5hJmBAalg6fhJjSOt0VEq2OCOO8aNi5FrJyS/OlcQ53Ow0W/fX9lPehIIwfUMC8alTFNocU6K+amPN8p5QZcvJsYBAtTESMlJH9PEHBIGgDmEmYEBqWDp+E/3Ju6O5fnm+JovHmoe10fkMC8xkYoRBnEd/VVeSjnhcKuyI4lPljksjRaQxz7DZVu/agf/y5MdcpqbYNE+u+kf08QcEgaAOYSZgQGpYOn4Ubd2+0bnC3AmH3wCpr9c+6IR2MnzId4A0IjbUkLSDCkSCMH1DAvGpUxTaHFOivmp3FCtlALoErAkoW3E+CMjBx/TxBwSBoA5hJmBAalg6fiWDonqkox8Hitk6UwNQH/ZOP1sTmOZEwIKQ0gAxLFPtRIIwfUMC8alTFNocU6K+an53ctiWXBCjuXZbYQY5YrLH9PEHBIGgDmEmYEBqWDp+DsNwW5i5v4+297UTUGk6aOkMq/uCOtHXJMWzXd582YahIk3qQjg0LgSOFGqzzOQKxv3WzYJrLaPLmkW0vtWS2Ef08QcEgaAOYSZgQGpYOn4Mu6gcRmP5W8pjDh8eTewwSckvzpXEOdzsNFv31/ZT3oSCMH1DAvGpUxTaHFOivmpWyrzt71xrONeeCOw+8T3mx/TxBwSBoA5hJmBAalg6fhvJ6NsnR3q4+XE9/F2KRKYJyS/OlcQ53Ow0W/fX9lPehIIwfUMC8alTFNocU6K+alcYquz1bLdJQuqbN8QlWgxH9PEHBIGgDmEmYEBqWDp+Eo6ukSDqgjjuZDKOWdYnFc4/WxOY5kTAgpDSADEsU+1EgjB9QwLxqVMU2hxTor5qYWHlj8hZYg4nDaH49HG3jQf08QcEgaAOYSZgQGpYOn4RyIRCv52pzsA9pKTbyFnuqr/ZeuBwRqNM/4PiodkEmO7OraWB16p6cJ1TEngcuXiFxvgAhp2CUjDtZIICn1Rxh/TxBwSBoA5hJmBAalg6fiTuhcC8pOMiuDs8bHYqEMmVe84UlBVIY7XZEWZ8XJBkxBpixGjXI1kf5luol5+eEZVI8K69XXozwMCBhCnBL/wH9PEHBIGgDmEmYEBqWDp+OvADpBAP6GNYQ8k33g5yS+iEdjJ8yHeANCI21JC0gwpEgjB9QwLxqVMU2hxTor5qUuMo68Gi7KWmbERE+eJaDYf08QcEgaAOYSZgQGpYOn4mKvSHnf32T8ov19P3Eyk2CckvzpXEOdzsNFv31/ZT3oSCMH1DAvGpUxTaHFOivmp44wUl7juS67tubmIHL5Hoh/TxBwSBoA5hJmBAalg6fj9JG0MFJmE3ZXNjMFy0CHZvq3g7+J4//u4z7Ab50QAI1d2Zt534RgWQyZUmDRQTkE1MtAkD5JllHam501uxiRsH9PEHBIGgDmEmYEBqWDp+L51yVaTijwyoobVjIjFXGXy2xC5khX3sY1wKKEfpGP2EgjB9QwLxqVMU2hxTor5qW40ibNYIvfSlrXWmbnaEdwf08QcEgaAOYSZgQGpYOn4nNYmwNZg1UMMLaWA/QP+gL6t4O/ieP/7uM+wG+dEACNXdmbed+EYFkMmVJg0UE5BrknhvT+q5Yhl9W+pzIZ4sx/TxBwSBoA5hJmBAalg6fjRVI9fuLLXvyOoPeR5mNiV8tsQuZIV97GNcCihH6Rj9hIIwfUMC8alTFNocU6K+anwSbnn3kbpxJtNE8uGhugGH9PEHBIGgDmEmYEBqWDp+PpF/cjo3vVkhKWqK2RSkuXTfpAXHkI010RLRiOR4MlYV3Zm3nfhGBZDJlSYNFBOQW+uCwgPpRYpCCXuBonhKOwf08QcEgaAOYSZgQGpYOn4U/Z8OcWqibRp88nIdhKbU2OGWZB7Quor6U7Ji2gYNukSCMH1DAvGpUxTaHFOivmpcLihLDS4O2o9hoc007WLQx/TxBwSBoA5hJmBAalg6fgSV98BayNefL7mlT4TTXe1aA4GPmSfXzsJ7dYEFhA4DPRjoT2cd5nIXHGBuunKpxATH+Nn5ULeoleHTFADwi65H9PEHBIGgDmEmYEBqWDp+JyWNLA2poaF98erzR8Cetc0Uuwm0uik5Ue+2ZZnc6AyVjH/LpFtIJkEm5KuK+k4+AxSQO3qFAroekFv2KRJlw0f08QcEgaAOYSZgQGpYOn4lyGfgvidpuuCygilRY2fPMdxo3Jqw9hnq96WL5pRRZcSCMH1DAvGpUxTaHFOivmpgHC48E44LxBqc4K0rdOyOx/TxBwSBoA5hJmBAalg6fjyBdPQYzrnIenb5bfX6tYX8tsQuZIV97GNcCihH6Rj9hIIwfUMC8alTFNocU6K+akZ7kBQ1NwsCZbDvVxcgih5H9PEHBIGgDmEmYEBqWDp+Acf4cjuqbW8SyHrQd0OVb+iEdjJ8yHeANCI21JC0gwpEgjB9QwLxqVMU2hxTor5qfc/alJsE2joWaRMDIUNZkkf08QcEgaAOYSZgQGpYOn4bAhhN1QeK+ywrvRoL5lk0CckvzpXEOdzsNFv31/ZT3oSCMH1DAvGpUxTaHFOivmpN1ejVkIMmCea+MgnS5Xnnx/TxBwSBoA5hJmBAalg6fjASKy33n3gxNg9wOfk3x0lY4ZZkHtC6ivpTsmLaBg26RIIwfUMC8alTFNocU6K+anPqc55WZIcTAKHsFU/Av9PH9PEHBIGgDmEmYEBqWDp+DYIj6tgzjOKODoAP26ZzSn4V0hOZAvicWWSnHKxYiHpV3Zm3nfhGBZDJlSYNFBOQf6/dQr/8TzVMNfvU8buGWgf08QcEgaAOYSZgQGpYOn4Mw29DL7kasPmvjSGa5chQ2OGWZB7Quor6U7Ji2gYNukSCMH1DAvGpUxTaHFOivmpEVVTzGtSACted8UBKa6Fux/TxBwSBoA5hJmBAalg6fiuI2bI7tHBWvHwhi6jtWSk2r13erpF+dSnkMkV7/PguxIIwfUMC8alTFNocU6K+alBsXdeCeEy88KLiIl0uQtLH9PEHBIGgDmEmYEBqWDp+GfRP+5uwgtUDF0+d9xqiKf4V0hOZAvicWWSnHKxYiHpV3Zm3nfhGBZDJlSYNFBOQQ5t+CaHqPtwKiHNV8YZOPkf08QcEgaAOYSZgQGpYOn4k9+AQek1xVK9mjfRJ1HmC+U8R31LKcQ/w4AFLN/JJTASCMH1DAvGpUxTaHFOivmps7rke53BsWqPqGoENIerbR/TxBwSBoA5hJmBAalg6fja4mj24zucDMCHE3bZ8X0gohHYyfMh3gDQiNtSQtIMKRIIwfUMC8alTFNocU6K+akB2jtbs0Pnd8zlKjQXC7MyH9PEHBIGgDmEmYEBqWDp+IbLnrjadFCbb+qW8YHAyx5H3wqv3wyq5Yrl8PbUij+3V3Zm3nfhGBZDJlSYNFBOQSedUx43MhSp8p0Ts+E0vwkf08QcEgaAOYSZgQGpYOn4QC41mzDUae7PGOc1UqguA0ffCq/fDKrliuXw9tSKP7dXdmbed+EYFkMmVJg0UE5BXr3Xzsjb12fOZOJum5ZcyR/TxBwSBoA5hJmBAalg6fhn5LBPmksGyKs3ue1/Wy+gxZFiuWgqKMTwlSVc+w+bn4KYayF+X3onpdWm8JzGxjvBp+9kU7BtRwacfGguk7BgH9PEHBIGgDmEmYEBqWDp+EciEQr+dqc7APaSk28hZ7outhmmLQVrPblwXnxVTFaW17SvoyZpa6+fyZv93e88ypgkMhZvCzwmTMUPGN3nRRAf08QcEgaAOYSZgQGpYOn4xVMim6/A0SpXr703tOg5g4xXqZjrgiLycXggrLvK0GuCmGshfl96J6XVpvCcxsY7cHSWZL5zuqj8K7NVDZOUvx/TxBwSBoA5hJmBAalg6fg+0bN/WLrhKWqemDnziNJcaD6WgwomxjRrZFV32wZy2qhXlgrpD7TLqHA2z09HkaIaudq69bdXxT5ZndK09DatH9PEHBIGgDmEmYEBqWDp+BYQAcUF+13XAUfGXZyiZDQ+++7JTdGH5Zmvf8OJr6wrAxBCQqQ6GD0PJYrp5eg501yO/HQX7KbMbfUx9o6bmzkf08QcEgaAOYSZgQGpYOn4ZAIy2nycs3ymZVVmYMI566IR2MnzId4A0IjbUkLSDCkSCMH1DAvGpUxTaHFOivmp82xSnwuf7egZlzRJ8oVOJx/TxBwSBoA5hJmBAalg6fiN/vK6lKKPhwoRG9fbnFzCrkmvul111U1SNzwlX+9qpK0rwiIYWzsCTct59Ky9aQfC0ooOlMjjbntCx9N+MLNnH9PEHBIGgDmEmYEBqWDp+OZ/YmP3iodn7u/ezlWXKONdXVyl5XkBpuFL8ts6ciF6K3tiqSoJkLf4PBmccc7ECniHBqA7vwhm5nEHuTUQjFD9jQvdIaEvNeRqPXi6kUkiOPE9faWIl/hjy2eyGB1KVE+ETfaU9UTnO3rv6C15dN2bIExjxqvAbyy5E+i81K0E6rSybYkivFPsFMWyx2kagP2NC90hoS815Go9eLqRSSIj5Qgo8XER6KNGhK6mvVC6Y4ZZkHtC6ivpTsmLaBg26RIIwfUMC8alTFNocU6K+alK/sFynN6Vn2GQf2QASp8l/Y0L3SGhLzXkaj14upFJImOqra5JIlgcyt/5qzBX4x1jhlmQe0LqK+lOyYtoGDbpEgjB9QwLxqVMU2hxTor5qYRrS5Mi5ZJA8IWU4OjsRMP9jQvdIaEvNeRqPXi6kUki9GxVVI96sAaUZK+yMbipe5pq2X09R/Amhz+F67g6AxybIExjxqvAbyy5E+i81K0EYQjrmvqzXAhZrFk3pLdetv2NC90hoS815Go9eLqRSSKmRMx7TpyekyTFwPUmXjrnLn5fqwOrZ5nrOWW4sBCX9M1z3SQj90R4DLm90AAFmspNizwCE3V0X0QCGC7qrk57/Y0L3SGhLzXkaj14upFJIkngBLuxlZakahk6qsh0yvAadKYAARFyVvM7kQ6DoTER5J9/tyieL+qKg78Kj2vYTLpoQdreQ2uP+cfPpNfMZy/9jQvdIaEvNeRqPXi6kUki0AJo8edrJl1SkeZ4RQC7kYkfvnSQNQHM2WdhEowQ2O033FcLjTse3oPBOfQHcQ9rqiJYfhzyRX2klBRRvL8M5B/TxBwSBoA5hJmBAalg6fhXps5LDStEcd0cil/kLiU0JyS/OlcQ53Ow0W/fX9lPehIIwfUMC8alTFNocU6K+anGD8twxh7sDomA5Kz0PdgvH9PEHBIGgDmEmYEBqWDp+BRsSlub2Ul0S7c6FDAXno72EQ7mD6LRBdzgrBma7EDsmyBMY8arwG8suRPovNStBDyAc7k/g3AfTj+x7h/Shyz9jQvdIaEvNeRqPXi6kUki20VlAm4qY0su0eGKp3RCK2RuBD3ohuPg+c4bs7W8VaIqf+RfSsPtLq56+DVgXXo0 | non_process | grfstb f x akuqockoinbkeujh e agf u cmjbx zeubwrqnm qp iep wg deacnxdmbed ppf dqr gfrp iblnrjadfcbb wy gxzfiuwgqkmtwlsvc w ecieqr byqacuf oz alk amhz tyiel cfppnfmzy | 0 |
45,906 | 24,264,970,100 | IssuesEvent | 2022-09-28 04:56:32 | appsmithorg/appsmith | https://api.github.com/repos/appsmithorg/appsmith | opened | [Task]: Improvements to performance dashboard | Performance Task | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### SubTasks
- [ ] Connect the perf app to a internal private git repo [https://github.com/appsmithorg/performance-infra-app](https://github.com/appsmithorg/performance-infra-app)
- [ ] On changing the PR, and run id, update the URL parameters [https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204](https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204)
- [ ] For SD in table, show in this format, absolute value(%)
- [ ] In this [[pr](https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204)](https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204) one of the commit ID doesn't belong to the PR, investigate why. [IMPORTANT]
- [ ] We are not using the Repo name in queries, we should fix that.
- [ ] Above the table, add a line graph to show how a tests metrics are doing on the entire repo. Highlight current PR points.
- [ ] We should be able to switch between using median/min values
- [ ] Figure out a way to use the running, mean of median and min value from when the test was run. Right now we are using the latest value [Not a high priority]
- [ ] A function to generate a running mean,min, sd of mean, sd of min. Good watch [https://www.youtube.com/watch?v=ADHMOBiBcFg&t=196s](https://www.youtube.com/watch?v=ADHMOBiBcFg&t=196s)
- [ ] Option to edit the run value, with a record of who edited it.
- [ ] A single run ID can have multiple runs. We are not accounting for it, we should use gh_run_attempt in the query and UI
- [ ] Add a button to re run the perf test. For this we need to write a workflow or update the current workflow to run only perf tests. | True | [Task]: Improvements to performance dashboard - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### SubTasks
- [ ] Connect the perf app to a internal private git repo [https://github.com/appsmithorg/performance-infra-app](https://github.com/appsmithorg/performance-infra-app)
- [ ] On changing the PR, and run id, update the URL parameters [https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204](https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204)
- [ ] For SD in table, show in this format, absolute value(%)
- [ ] In this [[pr](https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204)](https://app.appsmith.com/app/performance-dashboard-beta/pr-details-6324722ad54cbb52f19ac261?runId=3105332017&pr=16204) one of the commit ID doesn't belong to the PR, investigate why. [IMPORTANT]
- [ ] We are not using the Repo name in queries, we should fix that.
- [ ] Above the table, add a line graph to show how a tests metrics are doing on the entire repo. Highlight current PR points.
- [ ] We should be able to switch between using median/min values
- [ ] Figure out a way to use the running, mean of median and min value from when the test was run. Right now we are using the latest value [Not a high priority]
- [ ] A function to generate a running mean,min, sd of mean, sd of min. Good watch [https://www.youtube.com/watch?v=ADHMOBiBcFg&t=196s](https://www.youtube.com/watch?v=ADHMOBiBcFg&t=196s)
- [ ] Option to edit the run value, with a record of who edited it.
- [ ] A single run ID can have multiple runs. We are not accounting for it, we should use gh_run_attempt in the query and UI
- [ ] Add a button to re run the perf test. For this we need to write a workflow or update the current workflow to run only perf tests. | non_process | improvements to performance dashboard is there an existing issue for this i have searched the existing issues subtasks connect the perf app to a internal private git repo on changing the pr and run id update the url parameters for sd in table show in this format absolute value in this one of the commit id doesn t belong to the pr investigate why we are not using the repo name in queries we should fix that above the table add a line graph to show how a tests metrics are doing on the entire repo highlight current pr points we should be able to switch between using median min values figure out a way to use the running mean of median and min value from when the test was run right now we are using the latest value a function to generate a running mean min sd of mean sd of min good watch option to edit the run value with a record of who edited it a single run id can have multiple runs we are not accounting for it we should use gh run attempt in the query and ui add a button to re run the perf test for this we need to write a workflow or update the current workflow to run only perf tests | 0 |
7,631 | 10,730,486,773 | IssuesEvent | 2019-10-28 17:30:43 | prisma/prisma2 | https://api.github.com/repos/prisma/prisma2 | closed | Query engine: Replace HTTP server with another solution | process/candidate | ## Current problems
1. Severe Security problem: HTTP server is accessible from host machine and has privileged access to database 🚨
2. Shows misleading port detection messages on Codesandbox

## Solution
### Inherent solution
Get rid of HTTP server and come up with another solution (maybe similar to migration engine?).
### Temporary solution
Re 1) Secure HTTP endpoint with a auto-generated secret
Re 2) Use allowlist for HTTP ports on CSB
| 1.0 | Query engine: Replace HTTP server with another solution - ## Current problems
1. Severe Security problem: HTTP server is accessible from host machine and has privileged access to database 🚨
2. Shows misleading port detection messages on Codesandbox

## Solution
### Inherent solution
Get rid of HTTP server and come up with another solution (maybe similar to migration engine?).
### Temporary solution
Re 1) Secure HTTP endpoint with a auto-generated secret
Re 2) Use allowlist for HTTP ports on CSB
| process | query engine replace http server with another solution current problems severe security problem http server is accessible from host machine and has privileged access to database 🚨 shows misleading port detection messages on codesandbox solution inherent solution get rid of http server and come up with another solution maybe similar to migration engine temporary solution re secure http endpoint with a auto generated secret re use allowlist for http ports on csb | 1 |
257,239 | 22,153,249,698 | IssuesEvent | 2022-06-03 19:18:58 | apache/beam | https://api.github.com/repos/apache/beam | opened | Nexmark: migrate WinningBids to state API in place of custom window merge. | P3 improvement testing-nexmark | Custom window merge in WinningBids is expensive (builds a few maps). It would be simpler to use state API there.
Imported from Jira [BEAM-2971](https://issues.apache.org/jira/browse/BEAM-2971). Original Jira may contain additional context.
Reported by: echauchot. | 1.0 | Nexmark: migrate WinningBids to state API in place of custom window merge. - Custom window merge in WinningBids is expensive (builds a few maps). It would be simpler to use state API there.
Imported from Jira [BEAM-2971](https://issues.apache.org/jira/browse/BEAM-2971). Original Jira may contain additional context.
Reported by: echauchot. | non_process | nexmark migrate winningbids to state api in place of custom window merge custom window merge in winningbids is expensive builds a few maps it would be simpler to use state api there imported from jira original jira may contain additional context reported by echauchot | 0 |
233,416 | 25,765,442,170 | IssuesEvent | 2022-12-09 01:11:37 | dmyers87/prefect | https://api.github.com/repos/dmyers87/prefect | opened | CVE-2022-23491 (Medium) detected in certifi-2020.12.5-py2.py3-none-any.whl | security vulnerability | ## CVE-2022-23491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>certifi-2020.12.5-py2.py3-none-any.whl</b></p></summary>
<p>Python package for providing Mozilla's CA Bundle.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/5e/a0/5f06e1e1d463903cf0c0eebeb751791119ed7a4b3737fdc9a77f1cdfb51f/certifi-2020.12.5-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/5e/a0/5f06e1e1d463903cf0c0eebeb751791119ed7a4b3737fdc9a77f1cdfb51f/certifi-2020.12.5-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- docker-4.4.4-py2.py3-none-any.whl (Root Library)
- requests-2.25.1-py2.py3-none-any.whl
- :x: **certifi-2020.12.5-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Certifi is a curated collection of Root Certificates for validating the trustworthiness of SSL certificates while verifying the identity of TLS hosts. Certifi 2022.12.07 removes root certificates from "TrustCor" from the root store. These are in the process of being removed from Mozilla's trust store. TrustCor's root certificates are being removed pursuant to an investigation prompted by media reporting that TrustCor's ownership also operated a business that produced spyware. Conclusions of Mozilla's investigation can be found in the linked google group discussion.
<p>Publish Date: 2022-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23491>CVE-2022-23491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23491">https://www.cve.org/CVERecord?id=CVE-2022-23491</a></p>
<p>Release Date: 2022-12-07</p>
<p>Fix Resolution: certifi - 2022.12.07</p>
</p>
</details>
<p></p>
| True | CVE-2022-23491 (Medium) detected in certifi-2020.12.5-py2.py3-none-any.whl - ## CVE-2022-23491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>certifi-2020.12.5-py2.py3-none-any.whl</b></p></summary>
<p>Python package for providing Mozilla's CA Bundle.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/5e/a0/5f06e1e1d463903cf0c0eebeb751791119ed7a4b3737fdc9a77f1cdfb51f/certifi-2020.12.5-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/5e/a0/5f06e1e1d463903cf0c0eebeb751791119ed7a4b3737fdc9a77f1cdfb51f/certifi-2020.12.5-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- docker-4.4.4-py2.py3-none-any.whl (Root Library)
- requests-2.25.1-py2.py3-none-any.whl
- :x: **certifi-2020.12.5-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Certifi is a curated collection of Root Certificates for validating the trustworthiness of SSL certificates while verifying the identity of TLS hosts. Certifi 2022.12.07 removes root certificates from "TrustCor" from the root store. These are in the process of being removed from Mozilla's trust store. TrustCor's root certificates are being removed pursuant to an investigation prompted by media reporting that TrustCor's ownership also operated a business that produced spyware. Conclusions of Mozilla's investigation can be found in the linked google group discussion.
<p>Publish Date: 2022-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23491>CVE-2022-23491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23491">https://www.cve.org/CVERecord?id=CVE-2022-23491</a></p>
<p>Release Date: 2022-12-07</p>
<p>Fix Resolution: certifi - 2022.12.07</p>
</p>
</details>
<p></p>
| non_process | cve medium detected in certifi none any whl cve medium severity vulnerability vulnerable library certifi none any whl python package for providing mozilla s ca bundle library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy docker none any whl root library requests none any whl x certifi none any whl vulnerable library found in base branch master vulnerability details certifi is a curated collection of root certificates for validating the trustworthiness of ssl certificates while verifying the identity of tls hosts certifi removes root certificates from trustcor from the root store these are in the process of being removed from mozilla s trust store trustcor s root certificates are being removed pursuant to an investigation prompted by media reporting that trustcor s ownership also operated a business that produced spyware conclusions of mozilla s investigation can be found in the linked google group discussion publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope changed impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution certifi | 0 |
1,552 | 4,155,935,968 | IssuesEvent | 2016-06-16 16:20:02 | altoxml/schema | https://api.github.com/repos/altoxml/schema | closed | Provenance for OCRProcessing/Processing and Content | 1 submitted processing history | The current OCRProcessing statement is rather rudimentary in not allowing identifiers for each ProcessingStep and being able to link features in the recognition results to particular steps. For example, in our pipeline we frequently use tesseract's page segmentation with ocropus's recognition, so TextLine elements are sourced from one ProcessingStep and their text content is from another one.
A particular use case is when postprocessing like spell checkers add additional variants to String tags (something we'd like to see also) and it may be unclear if the variant is produced by the recognition engine itself or the spell checker. | 1.0 | Provenance for OCRProcessing/Processing and Content - The current OCRProcessing statement is rather rudimentary in not allowing identifiers for each ProcessingStep and being able to link features in the recognition results to particular steps. For example, in our pipeline we frequently use tesseract's page segmentation with ocropus's recognition, so TextLine elements are sourced from one ProcessingStep and their text content is from another one.
A particular use case is when postprocessing like spell checkers add additional variants to String tags (something we'd like to see also) and it may be unclear if the variant is produced by the recognition engine itself or the spell checker. | process | provenance for ocrprocessing processing and content the current ocrprocessing statement is rather rudimentary in not allowing identifiers for each processingstep and being able to link features in the recognition results to particular steps for example in our pipeline we frequently use tesseract s page segmentation with ocropus s recognition so textline elements are sourced from one processingstep and their text content is from another one a particular use case is when postprocessing like spell checkers add additional variants to string tags something we d like to see also and it may be unclear if the variant is produced by the recognition engine itself or the spell checker | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.