Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
113,244
| 17,116,224,153
|
IssuesEvent
|
2021-07-11 12:10:03
|
theHinneh/ha
|
https://api.github.com/repos/theHinneh/ha
|
closed
|
CVE-2019-1010266 (Medium) detected in lodash-3.10.1.tgz
|
security vulnerability
|
## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: ha/backend/package.json</p>
<p>Path to vulnerable library: ha/backend/node_modules/ioredis/node_modules/lodash/package.json,ha/backend/node_modules/kafka-node/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- mosca-2.8.3.tgz (Root Library)
- ioredis-1.15.1.tgz
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/theHinneh/ha/commit/b67d33dd9df9e05b70466e310843976220230240">b67d33dd9df9e05b70466e310843976220230240</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2019-07-17</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-1010266 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: ha/backend/package.json</p>
<p>Path to vulnerable library: ha/backend/node_modules/ioredis/node_modules/lodash/package.json,ha/backend/node_modules/kafka-node/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- mosca-2.8.3.tgz (Root Library)
- ioredis-1.15.1.tgz
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/theHinneh/ha/commit/b67d33dd9df9e05b70466e310843976220230240">b67d33dd9df9e05b70466e310843976220230240</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2019-07-17</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file ha backend package json path to vulnerable library ha backend node modules ioredis node modules lodash package json ha backend node modules kafka node node modules lodash package json dependency hierarchy mosca tgz root library ioredis tgz x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
328,578
| 28,125,935,891
|
IssuesEvent
|
2023-03-31 17:42:55
|
MicrosoftDocs/visualstudio-docs
|
https://api.github.com/repos/MicrosoftDocs/visualstudio-docs
|
closed
|
How to target an Azure cloud PC / Windows365 that I have access to
|
doc-bug visual-studio-windows/prod vs-ide-test/tech Pri2
|
How to target an Azure cloud PC / Windows365 that I have access to via Remote Desktop app.
I don't see ways to target the machine, as I have no clue what would be the public name, as it seems to be behind a RDP gateway. Is there ways to proxy the connection via that?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 90ec9455-ef64-8854-94bc-9be28887a6a4
* Version Independent ID: 9adeea5e-7ba1-065f-5de0-aae9147f9c85
* Content: [Remote Testing in Visual Studio - Visual Studio (Windows)](https://learn.microsoft.com/en-us/visualstudio/test/remote-testing?view=vs-2022)
* Content Source: [docs/test/remote-testing.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/main/docs/test/remote-testing.md)
* Product: **visual-studio-windows**
* Technology: **vs-ide-test**
* GitHub Login: @Mikejo5000
* Microsoft Alias: **mikejo**
|
1.0
|
How to target an Azure cloud PC / Windows365 that I have access to - How to target an Azure cloud PC / Windows365 that I have access to via Remote Desktop app.
I don't see ways to target the machine, as I have no clue what would be the public name, as it seems to be behind a RDP gateway. Is there ways to proxy the connection via that?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 90ec9455-ef64-8854-94bc-9be28887a6a4
* Version Independent ID: 9adeea5e-7ba1-065f-5de0-aae9147f9c85
* Content: [Remote Testing in Visual Studio - Visual Studio (Windows)](https://learn.microsoft.com/en-us/visualstudio/test/remote-testing?view=vs-2022)
* Content Source: [docs/test/remote-testing.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/main/docs/test/remote-testing.md)
* Product: **visual-studio-windows**
* Technology: **vs-ide-test**
* GitHub Login: @Mikejo5000
* Microsoft Alias: **mikejo**
|
non_process
|
how to target an azure cloud pc that i have access to how to target an azure cloud pc that i have access to via remote desktop app i don t see ways to target the machine as i have no clue what would be the public name as it seems to be behind a rdp gateway is there ways to proxy the connection via that document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product visual studio windows technology vs ide test github login microsoft alias mikejo
| 0
|
178,296
| 13,772,140,416
|
IssuesEvent
|
2020-10-07 23:41:53
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
closed
|
Schedule is not marked as a mandatory field in a CronJob create page
|
[zube]: To Test kind/bug
|
**On master-head - commit id: `33f755f18`**
- Workloads --> Cron Job --> create
- Schedule is not marked as a mandatory field in a CronJob create page
<img width="950" alt="Screen Shot 2020-09-17 at 3 03 21 PM" src="https://user-images.githubusercontent.com/26032343/93533061-0dc75800-f8f7-11ea-9d0c-0a23c18838da.png">
**Expected Result:**
Cron Job should be marked as a mandatory field
|
1.0
|
Schedule is not marked as a mandatory field in a CronJob create page - **On master-head - commit id: `33f755f18`**
- Workloads --> Cron Job --> create
- Schedule is not marked as a mandatory field in a CronJob create page
<img width="950" alt="Screen Shot 2020-09-17 at 3 03 21 PM" src="https://user-images.githubusercontent.com/26032343/93533061-0dc75800-f8f7-11ea-9d0c-0a23c18838da.png">
**Expected Result:**
Cron Job should be marked as a mandatory field
|
non_process
|
schedule is not marked as a mandatory field in a cronjob create page on master head commit id workloads cron job create schedule is not marked as a mandatory field in a cronjob create page img width alt screen shot at pm src expected result cron job should be marked as a mandatory field
| 0
|
13,513
| 3,343,449,948
|
IssuesEvent
|
2015-11-15 14:24:26
|
bolt/bolt
|
https://api.github.com/repos/bolt/bolt
|
closed
|
Multiple relation fields don't clear when single relation used
|
Blocking release Bug Needs Acceptance Test Needs Unit Test Regression
|
The value(s) fetched in `Bolt\Storage\Field\Type\RelationType::persist()` become `EntityProxy`
|
2.0
|
Multiple relation fields don't clear when single relation used - The value(s) fetched in `Bolt\Storage\Field\Type\RelationType::persist()` become `EntityProxy`
|
non_process
|
multiple relation fields don t clear when single relation used the value s fetched in bolt storage field type relationtype persist become entityproxy
| 0
|
93,588
| 3,906,046,838
|
IssuesEvent
|
2016-04-19 07:04:24
|
Captianrock/android_PV
|
https://api.github.com/repos/Captianrock/android_PV
|
opened
|
Dynamic updates for apps with traces
|
High Priority
|
Dynamically update the list of apps that have been analyzed by the user.
|
1.0
|
Dynamic updates for apps with traces - Dynamically update the list of apps that have been analyzed by the user.
|
non_process
|
dynamic updates for apps with traces dynamically update the list of apps that have been analyzed by the user
| 0
|
733,295
| 25,299,478,315
|
IssuesEvent
|
2022-11-17 09:38:06
|
opendatahub-io/odh-dashboard
|
https://api.github.com/repos/opendatahub-io/odh-dashboard
|
closed
|
[Model Serving]: Support Model Creation in Global View
|
kind/enhancement priority/blocker feature/model-serving
|
### Feature description
Follow up #648
Add the ability to deploy a model in the global view ([Mocks](https://www.sketch.com/s/113593f8-5970-49d6-a352-709b07639127/a/EL8AWlg))
### Describe alternatives you've considered
_No response_
### Anything else?
_No response_
|
1.0
|
[Model Serving]: Support Model Creation in Global View - ### Feature description
Follow up #648
Add the ability to deploy a model in the global view ([Mocks](https://www.sketch.com/s/113593f8-5970-49d6-a352-709b07639127/a/EL8AWlg))
### Describe alternatives you've considered
_No response_
### Anything else?
_No response_
|
non_process
|
support model creation in global view feature description follow up add the ability to deploy a model in the global view describe alternatives you ve considered no response anything else no response
| 0
|
16,661
| 21,728,177,662
|
IssuesEvent
|
2022-05-11 09:32:33
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Extract RestClient creation logic out of the ElasticsearchClient
|
kind/toil team/process-automation area/maintainability
|
**Description**
The `ElasticsearchClient` class creates a new `RestClient` based on user configuration passed via `ElasticsearchExporterConfiguration`. Splitting this out will allow us to test that the rest clients are properly constructed, and also allow us to reuse the same logic to construct high level REST clients for testing.
|
1.0
|
Extract RestClient creation logic out of the ElasticsearchClient - **Description**
The `ElasticsearchClient` class creates a new `RestClient` based on user configuration passed via `ElasticsearchExporterConfiguration`. Splitting this out will allow us to test that the rest clients are properly constructed, and also allow us to reuse the same logic to construct high level REST clients for testing.
|
process
|
extract restclient creation logic out of the elasticsearchclient description the elasticsearchclient class creates a new restclient based on user configuration passed via elasticsearchexporterconfiguration splitting this out will allow us to test that the rest clients are properly constructed and also allow us to reuse the same logic to construct high level rest clients for testing
| 1
|
4,720
| 7,552,846,759
|
IssuesEvent
|
2018-04-19 02:43:42
|
UnbFeelings/unb-feelings-docs
|
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
|
closed
|
[Não Conformidade] Relatório Final não existe
|
Processo Qualidade invalid
|
@UnbFeelings/process
Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o) fora auditada o artefato Relatório Final, resultante da atividade [Relatório Final de Medição](https://github.com/UnbFeelings/unb-feelings-docs/wiki/Processo#317-atividade-relatório-final-de-medição).
### Descrição
Foi identificado que não foram realizadas coletas de métricas de código fonte, ou se foram, não estão descritas em um artefato de acordo com o proposto pelo processo.
#### Recomendações
É recomendado a integração de ferramentas de análise de código para que as métricas possam ser geradas automaticamente, sem que haja a necessidade de atribuir esta tarefa a uma pessoa. No entanto recomenda-se definir um responsável pela elaboração do relatório de métricas.
#### Detalhes
**Auditor**: Jonathan Rufino
**Técnica de Audição**: Checklist
**Tipo:** Medição e Análise
**Prazo:** 23/04/2018
|
1.0
|
[Não Conformidade] Relatório Final não existe - @UnbFeelings/process
Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o) fora auditada o artefato Relatório Final, resultante da atividade [Relatório Final de Medição](https://github.com/UnbFeelings/unb-feelings-docs/wiki/Processo#317-atividade-relatório-final-de-medição).
### Descrição
Foi identificado que não foram realizadas coletas de métricas de código fonte, ou se foram, não estão descritas em um artefato de acordo com o proposto pelo processo.
#### Recomendações
É recomendado a integração de ferramentas de análise de código para que as métricas possam ser geradas automaticamente, sem que haja a necessidade de atribuir esta tarefa a uma pessoa. No entanto recomenda-se definir um responsável pela elaboração do relatório de métricas.
#### Detalhes
**Auditor**: Jonathan Rufino
**Técnica de Audição**: Checklist
**Tipo:** Medição e Análise
**Prazo:** 23/04/2018
|
process
|
relatório final não existe unbfeelings process perante critérios definidos para as fora auditada o artefato relatório final resultante da atividade descrição foi identificado que não foram realizadas coletas de métricas de código fonte ou se foram não estão descritas em um artefato de acordo com o proposto pelo processo recomendações é recomendado a integração de ferramentas de análise de código para que as métricas possam ser geradas automaticamente sem que haja a necessidade de atribuir esta tarefa a uma pessoa no entanto recomenda se definir um responsável pela elaboração do relatório de métricas detalhes auditor jonathan rufino técnica de audição checklist tipo medição e análise prazo
| 1
|
3,990
| 6,918,318,910
|
IssuesEvent
|
2017-11-29 11:43:50
|
nlbdev/pipeline
|
https://api.github.com/repos/nlbdev/pipeline
|
closed
|
Acrynyms with genitive "s" / Akronymer med genitivs-s
|
enhancement pre-processing Priority:1 - Low
|
(norwegian)
*from Trello-board (@matskober):*
Legge inn akronymer med genitivs-s (f.eks. NLBs. I punktskrift skal det være et tegn som markerer skillet mellom akronymet og genitivs-s - 56). Liste over akronymer er skaffet fra Mari
|
1.0
|
Acrynyms with genitive "s" / Akronymer med genitivs-s - (norwegian)
*from Trello-board (@matskober):*
Legge inn akronymer med genitivs-s (f.eks. NLBs. I punktskrift skal det være et tegn som markerer skillet mellom akronymet og genitivs-s - 56). Liste over akronymer er skaffet fra Mari
|
process
|
acrynyms with genitive s akronymer med genitivs s norwegian from trello board matskober legge inn akronymer med genitivs s f eks nlbs i punktskrift skal det være et tegn som markerer skillet mellom akronymet og genitivs s liste over akronymer er skaffet fra mari
| 1
|
13,650
| 8,306,928,655
|
IssuesEvent
|
2018-09-23 01:11:10
|
VSCodeVim/Vim
|
https://api.github.com/repos/VSCodeVim/Vim
|
closed
|
Investigate reducing startup activation time
|
area/performance
|
- Extension Name: vim
- Extension Version: 0.15.7
- OS Version: Windows_NT x64 10.0.15063
- VSCode version: 1.25.1
We have written the needed data into your clipboard. Please paste:

why soooooooooooooooooooo slow???!!!!
|
True
|
Investigate reducing startup activation time - - Extension Name: vim
- Extension Version: 0.15.7
- OS Version: Windows_NT x64 10.0.15063
- VSCode version: 1.25.1
We have written the needed data into your clipboard. Please paste:

why soooooooooooooooooooo slow???!!!!
|
non_process
|
investigate reducing startup activation time extension name vim extension version os version windows nt vscode version we have written the needed data into your clipboard please paste why soooooooooooooooooooo slow???!!!!
| 0
|
218,628
| 16,766,016,398
|
IssuesEvent
|
2021-06-14 08:54:06
|
jakobbossek/ecr3vis
|
https://api.github.com/repos/jakobbossek/ecr3vis
|
opened
|
Use mathjax for HTML output
|
documentation
|
We have many math formulas in the documentation. These are not rendered in the HTML RD files. Consider using [mathjaxr](http://cran.uni-muenster.de/web/packages/mathjaxr/mathjaxr.pdf)
|
1.0
|
Use mathjax for HTML output - We have many math formulas in the documentation. These are not rendered in the HTML RD files. Consider using [mathjaxr](http://cran.uni-muenster.de/web/packages/mathjaxr/mathjaxr.pdf)
|
non_process
|
use mathjax for html output we have many math formulas in the documentation these are not rendered in the html rd files consider using
| 0
|
349,799
| 31,831,824,270
|
IssuesEvent
|
2023-09-14 11:06:06
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachprod: implement "Reset" for aws and azure
|
C-enhancement X-stale A-roachprod no-issue-activity T-testeng O-cloudreport
|
Only GCE supports `Reset`--restarting a VM. As a result, some roachtests may be flaky (or incorrect) when executing outside of GCE. E.g., tpcc roachperf uses `Reset` [1] during each iteration of line search to determine an optimal number of warehouses. While the restart after each iteration is technically not required, it reduces noise.
Other examples include failure-injection scenarios, e.g., restarting a VM at random.
[1] https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/tests/tpcc.go#L1111
Epic: CRDB-10428
Jira issue: CRDB-13686
|
1.0
|
roachprod: implement "Reset" for aws and azure - Only GCE supports `Reset`--restarting a VM. As a result, some roachtests may be flaky (or incorrect) when executing outside of GCE. E.g., tpcc roachperf uses `Reset` [1] during each iteration of line search to determine an optimal number of warehouses. While the restart after each iteration is technically not required, it reduces noise.
Other examples include failure-injection scenarios, e.g., restarting a VM at random.
[1] https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/tests/tpcc.go#L1111
Epic: CRDB-10428
Jira issue: CRDB-13686
|
non_process
|
roachprod implement reset for aws and azure only gce supports reset restarting a vm as a result some roachtests may be flaky or incorrect when executing outside of gce e g tpcc roachperf uses reset during each iteration of line search to determine an optimal number of warehouses while the restart after each iteration is technically not required it reduces noise other examples include failure injection scenarios e g restarting a vm at random epic crdb jira issue crdb
| 0
|
19,448
| 25,727,174,600
|
IssuesEvent
|
2022-12-07 17:26:47
|
RobertCraigie/prisma-client-py
|
https://api.github.com/repos/RobertCraigie/prisma-client-py
|
opened
|
Could not connect to the Query Engine due to OSError [Errno 99]
|
bug/2-confirmed kind/bug process/candidate level/intermediate priority/high
|
<!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
A user has encountered this error:
```
OSError: [Errno 99] Cannot assign requested address
File "httpcore/_exceptions.py", line 10, in map_exceptions
yield
File "httpcore/backends/sync.py", line 94, in connect_tcp
sock = socket.create_connection(
File "socket.py", line 844, in create_connection
raise err
File "socket.py", line 832, in create_connection
sock.connect(sa)
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
Not currently reproducible. Theorized cause is a race condition.
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
This should not crash.
|
1.0
|
Could not connect to the Query Engine due to OSError [Errno 99] - <!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
A user has encountered this error:
```
OSError: [Errno 99] Cannot assign requested address
File "httpcore/_exceptions.py", line 10, in map_exceptions
yield
File "httpcore/backends/sync.py", line 94, in connect_tcp
sock = socket.create_connection(
File "socket.py", line 844, in create_connection
raise err
File "socket.py", line 832, in create_connection
sock.connect(sa)
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
Not currently reproducible. Theorized cause is a race condition.
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
This should not crash.
|
process
|
could not connect to the query engine due to oserror thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description a user has encountered this error oserror cannot assign requested address file httpcore exceptions py line in map exceptions yield file httpcore backends sync py line in connect tcp sock socket create connection file socket py line in create connection raise err file socket py line in create connection sock connect sa how to reproduce steps to reproduce the behavior go to change run see error not currently reproducible theorized cause is a race condition expected behavior this should not crash
| 1
|
834
| 3,297,247,142
|
IssuesEvent
|
2015-11-02 07:26:24
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
System.Diagnostics.Process missing environment variables.
|
System.Diagnostics.Process
|
/cc @Priya91, @pallavit, @joshfree, @stephentoub
I am using Process to run a "Build.cmd", I have `echo %PATH%`, The result which run by C# is empty, at the same time, I tried add the system env vars again, it still not works. However, I double click the "Build.cmd", It can run the correct %PATH%
```
Process = new Process();
WorkingDirectory = FindDirectory(workingDirectory);
var fileName = "cmd.exe";
if (OS.Current != OSType.Windows)
{
fileName = "bash";
}
var arguments = "/c build.cmd";
if (OS.Current != OSType.Windows)
{
arguments = "./build.sh";
}
Process.StartInfo = new ProcessStartInfo
{
FileName = fileName,
Arguments = arguments,
UseShellExecute = false,
RedirectStandardError = true,
RedirectStandardOutput = true,
RedirectStandardInput = true,
WorkingDirectory = WorkingDirectory
};
```
...
```
var sysenv = Environment.GetEnvironmentVariables();
foreach(dynamic ev in sysenv)
{
#if DNXCORE50
if (Process.StartInfo.Environment[ev.Key] != null)
Process.StartInfo.Environment[ev.Key] = Process.StartInfo.Environment[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value;
else
Process.StartInfo.Environment.Add(ev.Key, ev.Value);
#else
if (Process.StartInfo.EnvironmentVariables[ev.Key] != null)
Process.StartInfo.EnvironmentVariables[ev.Key] = Process.StartInfo.EnvironmentVariables[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value;
else
Process.StartInfo.EnvironmentVariables.Add(ev.Key, ev.Value);
#endif
}
```
|
1.0
|
System.Diagnostics.Process missing environment variables. - /cc @Priya91, @pallavit, @joshfree, @stephentoub
I am using Process to run a "Build.cmd", I have `echo %PATH%`, The result which run by C# is empty, at the same time, I tried add the system env vars again, it still not works. However, I double click the "Build.cmd", It can run the correct %PATH%
```
Process = new Process();
WorkingDirectory = FindDirectory(workingDirectory);
var fileName = "cmd.exe";
if (OS.Current != OSType.Windows)
{
fileName = "bash";
}
var arguments = "/c build.cmd";
if (OS.Current != OSType.Windows)
{
arguments = "./build.sh";
}
Process.StartInfo = new ProcessStartInfo
{
FileName = fileName,
Arguments = arguments,
UseShellExecute = false,
RedirectStandardError = true,
RedirectStandardOutput = true,
RedirectStandardInput = true,
WorkingDirectory = WorkingDirectory
};
```
...
```
var sysenv = Environment.GetEnvironmentVariables();
foreach(dynamic ev in sysenv)
{
#if DNXCORE50
if (Process.StartInfo.Environment[ev.Key] != null)
Process.StartInfo.Environment[ev.Key] = Process.StartInfo.Environment[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value;
else
Process.StartInfo.Environment.Add(ev.Key, ev.Value);
#else
if (Process.StartInfo.EnvironmentVariables[ev.Key] != null)
Process.StartInfo.EnvironmentVariables[ev.Key] = Process.StartInfo.EnvironmentVariables[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value;
else
Process.StartInfo.EnvironmentVariables.Add(ev.Key, ev.Value);
#endif
}
```
|
process
|
system diagnostics process missing environment variables cc pallavit joshfree stephentoub i am using process to run a build cmd i have echo path the result which run by c is empty at the same time i tried add the system env vars again it still not works however i double click the build cmd it can run the correct path process new process workingdirectory finddirectory workingdirectory var filename cmd exe if os current ostype windows filename bash var arguments c build cmd if os current ostype windows arguments build sh process startinfo new processstartinfo filename filename arguments arguments useshellexecute false redirectstandarderror true redirectstandardoutput true redirectstandardinput true workingdirectory workingdirectory var sysenv environment getenvironmentvariables foreach dynamic ev in sysenv if if process startinfo environment null process startinfo environment process startinfo environment trimend trimend ev value else process startinfo environment add ev key ev value else if process startinfo environmentvariables null process startinfo environmentvariables process startinfo environmentvariables trimend trimend ev value else process startinfo environmentvariables add ev key ev value endif
| 1
|
14,234
| 17,154,611,056
|
IssuesEvent
|
2021-07-14 04:15:51
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
opened
|
Provision for app upgrades
|
Android P1 Participant datastore Process: Enhancement iOS
|
1. Notify app users if a new version of the mobile app is available on the app stores. The notification should redirect to the app store page.
2. Have a provision to configure forced vs. optional upgrade app behavior. If there is a new app version available, users see a message when they launch or visit the app, asking them to upgrade the app before continuing to use it. This can be a mandatory or optional step depending on a server-side configuration.
3. Users should be able to resume app usage smoothly post app upgrade. Requiring them to sign in again is acceptable if there are server-side changes that necessitate this.
|
1.0
|
Provision for app upgrades - 1. Notify app users if a new version of the mobile app is available on the app stores. The notification should redirect to the app store page.
2. Have a provision to configure forced vs. optional upgrade app behavior. If there is a new app version available, users see a message when they launch or visit the app, asking them to upgrade the app before continuing to use it. This can be a mandatory or optional step depending on a server-side configuration.
3. Users should be able to resume app usage smoothly post app upgrade. Requiring them to sign in again is acceptable if there are server-side changes that necessitate this.
|
process
|
provision for app upgrades notify app users if a new version of the mobile app is available on the app stores the notification should redirect to the app store page have a provision to configure forced vs optional upgrade app behavior if there is a new app version available users see a message when they launch or visit the app asking them to upgrade the app before continuing to use it this can be a mandatory or optional step depending on a server side configuration users should be able to resume app usage smoothly post app upgrade requiring them to sign in again is acceptable if there are server side changes that necessitate this
| 1
|
502,276
| 14,543,479,091
|
IssuesEvent
|
2020-12-15 16:55:44
|
zulip/zulip-mobile
|
https://api.github.com/repos/zulip/zulip-mobile
|
closed
|
Add mobile support for new `user_avatar_url_field_optional` client capability
|
P1 high-priority a-avatar api migrations
|
To resolve https://github.com/zulip/zulip/pull/15287, we're introducing a new client_capability that should allow the mobile app to have good performance when talking to servers with thousands of long-term-idle users (and email_address_visibility configured; that last bit being relevant mainly in that our previous attempt at solving this problem. `client_gravatar` feature works only with EMAIL_ADDRESS_VISIBILITY_EVERYONE), since the client needs real email addresses to compute gravatar hashes.
One should be able to prototype today against https://github.com/zulip/zulip/pull/15359; it should work aside from having the wrong capability name. I expect that to get cleaned up and this to merge in the next few days.
Tagging as a priority since this issue makes chat.zulip.org a lot slower to load on mobile.
|
1.0
|
Add mobile support for new `user_avatar_url_field_optional` client capability - To resolve https://github.com/zulip/zulip/pull/15287, we're introducing a new client_capability that should allow the mobile app to have good performance when talking to servers with thousands of long-term-idle users (and email_address_visibility configured; that last bit being relevant mainly in that our previous attempt at solving this problem. `client_gravatar` feature works only with EMAIL_ADDRESS_VISIBILITY_EVERYONE), since the client needs real email addresses to compute gravatar hashes.
One should be able to prototype today against https://github.com/zulip/zulip/pull/15359; it should work aside from having the wrong capability name. I expect that to get cleaned up and this to merge in the next few days.
Tagging as a priority since this issue makes chat.zulip.org a lot slower to load on mobile.
|
non_process
|
add mobile support for new user avatar url field optional client capability to resolve we re introducing a new client capability that should allow the mobile app to have good performance when talking to servers with thousands of long term idle users and email address visibility configured that last bit being relevant mainly in that our previous attempt at solving this problem client gravatar feature works only with email address visibility everyone since the client needs real email addresses to compute gravatar hashes one should be able to prototype today against it should work aside from having the wrong capability name i expect that to get cleaned up and this to merge in the next few days tagging as a priority since this issue makes chat zulip org a lot slower to load on mobile
| 0
|
71,943
| 9,545,095,416
|
IssuesEvent
|
2019-05-01 16:02:33
|
CosmiQ/cw-nets
|
https://api.github.com/repos/CosmiQ/cw-nets
|
closed
|
Augmentation docs
|
Difficulty: Easy Priority: High Status: On Hold Type: Documentation
|
After completing augmentation implementation (#35) we need to document it.
Documentation components:
[ ] list of available augmentations
[ ] set of augmentations only compatible with 3-channel imagery
[ ] set of required arguments for each augmentation
[ ] instructions and examples for using the `cw_nets.data.transform` submodule, including yaml config formatting
|
1.0
|
Augmentation docs - After completing augmentation implementation (#35) we need to document it.
Documentation components:
[ ] list of available augmentations
[ ] set of augmentations only compatible with 3-channel imagery
[ ] set of required arguments for each augmentation
[ ] instructions and examples for using the `cw_nets.data.transform` submodule, including yaml config formatting
|
non_process
|
augmentation docs after completing augmentation implementation we need to document it documentation components list of available augmentations set of augmentations only compatible with channel imagery set of required arguments for each augmentation instructions and examples for using the cw nets data transform submodule including yaml config formatting
| 0
|
7,359
| 10,509,141,261
|
IssuesEvent
|
2019-09-27 10:14:32
|
prisma/studio
|
https://api.github.com/repos/prisma/studio
|
opened
|
Reload unintuitive
|
kind/improvement process/candidate
|
## Reload in Table Plus

## Reload in Studio

As a user, who just wants the data to reload, without knowing that Photon is used under the hood, I don't know, what `Run` means in this context. I just want the UI to reload.
I suggest either calling it `Reload` or replacing it with a reload icon.
|
1.0
|
Reload unintuitive - ## Reload in Table Plus

## Reload in Studio

As a user, who just wants the data to reload, without knowing that Photon is used under the hood, I don't know, what `Run` means in this context. I just want the UI to reload.
I suggest either calling it `Reload` or replacing it with a reload icon.
|
process
|
reload unintuitive reload in table plus reload in studio as a user who just wants the data to reload without knowing that photon is used under the hood i don t know what run means in this context i just want the ui to reload i suggest either calling it reload or replacing it with a reload icon
| 1
|
3,233
| 6,289,280,115
|
IssuesEvent
|
2017-07-19 18:51:21
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess fails with NRE on UAP
|
area-System.Diagnostics.Process bug os-windows-uwp
|
(Test case will be added soon, creating issue so that I can disable that in the PR)
```
ERROR: System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess [FAIL]
System.AggregateException : One or more errors occurred. (Object reference not set to an instance of an object.) (Object reference not s
et to an instance of an object.)
---- System.NullReferenceException : Object reference not set to an instance of an object.
---- System.NullReferenceException : Object reference not set to an instance of an object.
Stack Trace:
----- Inner Stack Trace #1 (System.NullReferenceException) -----
at System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess()
----- Inner Stack Trace #2 (System.NullReferenceException) -----
at System.Diagnostics.Tests.ProcessTestBase.Dispose(Boolean disposing)
at System.IO.FileCleanupTestBase.Dispose()
at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, Can
cellationTokenSource cancellationTokenSource)
```
|
1.0
|
System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess fails with NRE on UAP - (Test case will be added soon, creating issue so that I can disable that in the PR)
```
ERROR: System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess [FAIL]
System.AggregateException : One or more errors occurred. (Object reference not set to an instance of an object.) (Object reference not s
et to an instance of an object.)
---- System.NullReferenceException : Object reference not set to an instance of an object.
---- System.NullReferenceException : Object reference not set to an instance of an object.
Stack Trace:
----- Inner Stack Trace #1 (System.NullReferenceException) -----
at System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess()
----- Inner Stack Trace #2 (System.NullReferenceException) -----
at System.Diagnostics.Tests.ProcessTestBase.Dispose(Boolean disposing)
at System.IO.FileCleanupTestBase.Dispose()
at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, Can
cellationTokenSource cancellationTokenSource)
```
|
process
|
system diagnostics tests processwaitingtests waitforpeerprocess fails with nre on uap test case will be added soon creating issue so that i can disable that in the pr error system diagnostics tests processwaitingtests waitforpeerprocess system aggregateexception one or more errors occurred object reference not set to an instance of an object object reference not s et to an instance of an object system nullreferenceexception object reference not set to an instance of an object system nullreferenceexception object reference not set to an instance of an object stack trace inner stack trace system nullreferenceexception at system diagnostics tests processwaitingtests waitforpeerprocess inner stack trace system nullreferenceexception at system diagnostics tests processtestbase dispose boolean disposing at system io filecleanuptestbase dispose at reflectionabstractionextensions disposetestclass itest test object testclass imessagebus messagebus executiontimer timer can cellationtokensource cancellationtokensource
| 1
|
125,894
| 17,861,285,972
|
IssuesEvent
|
2021-09-06 01:06:02
|
bsbtd/Teste
|
https://api.github.com/repos/bsbtd/Teste
|
opened
|
CVE-2021-23437 (High) detected in Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2021-23437 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: Teste/pytorch-metric-learning</p>
<p>Path to vulnerable library: Teste/pytorch-metric-learning</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package pillow from 0 and before 8.3.2 are vulnerable to Regular Expression Denial of Service (ReDoS) via the getrgb function.
<p>Publish Date: 2021-09-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23437>CVE-2021-23437</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html">https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html</a></p>
<p>Release Date: 2021-09-03</p>
<p>Fix Resolution: Pillow - 8.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23437 (High) detected in Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl - ## CVE-2021-23437 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: Teste/pytorch-metric-learning</p>
<p>Path to vulnerable library: Teste/pytorch-metric-learning</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package pillow from 0 and before 8.3.2 are vulnerable to Regular Expression Denial of Service (ReDoS) via the getrgb function.
<p>Publish Date: 2021-09-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23437>CVE-2021-23437</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html">https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html</a></p>
<p>Release Date: 2021-09-03</p>
<p>Fix Resolution: Pillow - 8.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file teste pytorch metric learning path to vulnerable library teste pytorch metric learning dependency hierarchy x pillow whl vulnerable library vulnerability details the package pillow from and before are vulnerable to regular expression denial of service redos via the getrgb function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pillow step up your open source security game with whitesource
| 0
|
22,537
| 31,707,807,504
|
IssuesEvent
|
2023-09-09 00:18:23
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
opened
|
Can’t scale up a node pool because of a failing scheduling predicate
|
bug process
|
### Description
Seen a couple times where GKE node pools can't scale because of a failing scheduling predicate. Logs indicate it was promtail and node exporter daemonset. These daemonset pods could not get scheduled because other items on the node had higher priority. This would be fine by itself since these pods are not the highest priority, but if it causes the node to not scale then it needs to be addressed.
### Steps to reproduce
Monitor GKE with real life workloads.
### Additional context
_No response_
### Hedera network
other
### Version
0.87.0
### Operating system
None
|
1.0
|
Can’t scale up a node pool because of a failing scheduling predicate - ### Description
Seen a couple times where GKE node pools can't scale because of a failing scheduling predicate. Logs indicate it was promtail and node exporter daemonset. These daemonset pods could not get scheduled because other items on the node had higher priority. This would be fine by itself since these pods are not the highest priority, but if it causes the node to not scale then it needs to be addressed.
### Steps to reproduce
Monitor GKE with real life workloads.
### Additional context
_No response_
### Hedera network
other
### Version
0.87.0
### Operating system
None
|
process
|
can’t scale up a node pool because of a failing scheduling predicate description seen a couple times where gke node pools can t scale because of a failing scheduling predicate logs indicate it was promtail and node exporter daemonset these daemonset pods could not get scheduled because other items on the node had higher priority this would be fine by itself since these pods are not the highest priority but if it causes the node to not scale then it needs to be addressed steps to reproduce monitor gke with real life workloads additional context no response hedera network other version operating system none
| 1
|
212,093
| 16,472,784,182
|
IssuesEvent
|
2021-05-23 18:53:41
|
truecharts/apps
|
https://api.github.com/repos/truecharts/apps
|
closed
|
Adapt for persitance.emptyDir to persistance.emptyDir.enabled
|
documentation enhancement good first issue
|
**Is your feature request related to a problem? Please describe.**
common 4.0.0 gave persistance.emptyDir it's own sub parameter `enabled`
**Describe the solution you'd like**
Adapt current docs and charts accordingly.
|
1.0
|
Adapt for persitance.emptyDir to persistance.emptyDir.enabled - **Is your feature request related to a problem? Please describe.**
common 4.0.0 gave persistance.emptyDir it's own sub parameter `enabled`
**Describe the solution you'd like**
Adapt current docs and charts accordingly.
|
non_process
|
adapt for persitance emptydir to persistance emptydir enabled is your feature request related to a problem please describe common gave persistance emptydir it s own sub parameter enabled describe the solution you d like adapt current docs and charts accordingly
| 0
|
3,714
| 6,732,600,623
|
IssuesEvent
|
2017-10-18 12:10:11
|
lockedata/rcms
|
https://api.github.com/repos/lockedata/rcms
|
opened
|
Manage attendees
|
conference team osem processes
|
## Detailed task
- Monitor sales
- Modify a registration e.g. issue a refund
- Send an email to attendees
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [osem](https://intense-shore-93790.herokuapp.com/)
- System documentation: [osem docs](http://osem.io/)
- Role: Conference team
- Area: Processes
|
1.0
|
Manage attendees - ## Detailed task
- Monitor sales
- Modify a registration e.g. issue a refund
- Send an email to attendees
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [osem](https://intense-shore-93790.herokuapp.com/)
- System documentation: [osem docs](http://osem.io/)
- Role: Conference team
- Area: Processes
|
process
|
manage attendees detailed task monitor sales modify a registration e g issue a refund send an email to attendees assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role conference team area processes
| 1
|
18,530
| 24,552,656,670
|
IssuesEvent
|
2022-10-12 13:43:18
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] [Offline indicator] My Account > Toggle buttons should not be functional in offline behaviour
|
Bug P1 iOS Process: Fixed Process: Tested dev
|
**Steps:**
1. Install the app
2. Sign in/signup
3. Navigated to My account
4. Switch off the internet
6. Now enable/disable toggle buttons like 'Receive push notifications?' or 'Receive study activity reminders?'
7. Switch on the internet
8. Navigate to different screen and come back to my account
9. Observe buttons enabled/disabled fails to retain
**Actual:** Button functionality fails to retain once navigated from different screen to my account
**Expected:** Toggle buttons should not be functional in offline behaviour
Refer video
https://user-images.githubusercontent.com/60386291/179733053-e6086ff4-ba51-42e4-86a2-6d3f3802a6e2.MOV
|
2.0
|
[iOS] [Offline indicator] My Account > Toggle buttons should not be functional in offline behaviour - **Steps:**
1. Install the app
2. Sign in/signup
3. Navigated to My account
4. Switch off the internet
6. Now enable/disable toggle buttons like 'Receive push notifications?' or 'Receive study activity reminders?'
7. Switch on the internet
8. Navigate to different screen and come back to my account
9. Observe buttons enabled/disabled fails to retain
**Actual:** Button functionality fails to retain once navigated from different screen to my account
**Expected:** Toggle buttons should not be functional in offline behaviour
Refer video
https://user-images.githubusercontent.com/60386291/179733053-e6086ff4-ba51-42e4-86a2-6d3f3802a6e2.MOV
|
process
|
my account toggle buttons should not be functional in offline behaviour steps install the app sign in signup navigated to my account switch off the internet now enable disable toggle buttons like receive push notifications or receive study activity reminders switch on the internet navigate to different screen and come back to my account observe buttons enabled disabled fails to retain actual button functionality fails to retain once navigated from different screen to my account expected toggle buttons should not be functional in offline behaviour refer video
| 1
|
70,963
| 13,564,458,780
|
IssuesEvent
|
2020-09-18 10:04:28
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
closed
|
Death desync after moving sub away from spawnpoint
|
Bug Code High prio Networking
|
- [X] I have searched the issue tracker to check if the issue has already been reported.
**Description**
~After respawning in a shuttle on a mission~ After moving away from spawnpoint there will be massive death-related desync. When you die server-side you will be locked in place instead. When you are knocked out you can't `give up` and spectate.
**Steps To Reproduce** (Original, see below for a non-shuttle related one, though this one works too)
1. Host server (dedicated or p2p).
2. Start mission.
3. Kill yourself (for example `kill` command).
4. Respawn in a shuttle (`respawnnow` command).
5. Kill yourself again (`kill` command).
6. Notice that you will die on the server, but you will still appear very alive for some time but can't move or give up.
**Version**
0.10.505.0 - 0.10.5
**Additional information**
I'm pretty sure it did not happen in the previous minor unstable version.
EDIT: Update description, version and steps to reproduce with newly found information.
|
1.0
|
Death desync after moving sub away from spawnpoint - - [X] I have searched the issue tracker to check if the issue has already been reported.
**Description**
~After respawning in a shuttle on a mission~ After moving away from spawnpoint there will be massive death-related desync. When you die server-side you will be locked in place instead. When you are knocked out you can't `give up` and spectate.
**Steps To Reproduce** (Original, see below for a non-shuttle related one, though this one works too)
1. Host server (dedicated or p2p).
2. Start mission.
3. Kill yourself (for example `kill` command).
4. Respawn in a shuttle (`respawnnow` command).
5. Kill yourself again (`kill` command).
6. Notice that you will die on the server, but you will still appear very alive for some time but can't move or give up.
**Version**
0.10.505.0 - 0.10.5
**Additional information**
I'm pretty sure it did not happen in the previous minor unstable version.
EDIT: Update description, version and steps to reproduce with newly found information.
|
non_process
|
death desync after moving sub away from spawnpoint i have searched the issue tracker to check if the issue has already been reported description after respawning in a shuttle on a mission after moving away from spawnpoint there will be massive death related desync when you die server side you will be locked in place instead when you are knocked out you can t give up and spectate steps to reproduce original see below for a non shuttle related one though this one works too host server dedicated or start mission kill yourself for example kill command respawn in a shuttle respawnnow command kill yourself again kill command notice that you will die on the server but you will still appear very alive for some time but can t move or give up version additional information i m pretty sure it did not happen in the previous minor unstable version edit update description version and steps to reproduce with newly found information
| 0
|
17,944
| 5,535,467,504
|
IssuesEvent
|
2017-03-21 17:25:46
|
phetsims/masses-and-springs
|
https://api.github.com/repos/phetsims/masses-and-springs
|
opened
|
Factor out duplicated code in line creation
|
dev:code-review
|
During #36 I saw this code in IndicatorVisibilityControlPanel:
```js
// Lines added for reference in panel
var greenLine = new Line( 0, 0, LINE_LENGTH, 0, {
stroke: 'rgb(93, 191, 142)',
lineDash: [ 6, 2.5 ],
lineWidth: 2.0,
cursor: 'pointer',
tandem: tandem.createTandem( 'greenLine' )
} );
var blueLine = new Line( 0, 0, LINE_LENGTH, 0, {
stroke: 'rgb(65,66,232)',
lineDash: [ 6, 2.5 ],
lineWidth: 2.0,
cursor: 'pointer',
tandem: tandem.createTandem( 'blueLine' )
} );
var redLine = new Line( 0, 0, LINE_LENGTH, 0, {
stroke: 'red',
lineDash: [ 6, 2.5 ],
lineWidth: 2.0,
cursor: 'pointer',
tandem: tandem.createTandem( 'redLine' )
} );
```
I recommend factoring out a function so the lines can be created like this:
```js
var greenLine = createLine('rgb(93, 191, 142)',tandem.createTandem('greenLine'));
```
|
1.0
|
Factor out duplicated code in line creation - During #36 I saw this code in IndicatorVisibilityControlPanel:
```js
// Lines added for reference in panel
var greenLine = new Line( 0, 0, LINE_LENGTH, 0, {
stroke: 'rgb(93, 191, 142)',
lineDash: [ 6, 2.5 ],
lineWidth: 2.0,
cursor: 'pointer',
tandem: tandem.createTandem( 'greenLine' )
} );
var blueLine = new Line( 0, 0, LINE_LENGTH, 0, {
stroke: 'rgb(65,66,232)',
lineDash: [ 6, 2.5 ],
lineWidth: 2.0,
cursor: 'pointer',
tandem: tandem.createTandem( 'blueLine' )
} );
var redLine = new Line( 0, 0, LINE_LENGTH, 0, {
stroke: 'red',
lineDash: [ 6, 2.5 ],
lineWidth: 2.0,
cursor: 'pointer',
tandem: tandem.createTandem( 'redLine' )
} );
```
I recommend factoring out a function so the lines can be created like this:
```js
var greenLine = createLine('rgb(93, 191, 142)',tandem.createTandem('greenLine'));
```
|
non_process
|
factor out duplicated code in line creation during i saw this code in indicatorvisibilitycontrolpanel js lines added for reference in panel var greenline new line line length stroke rgb linedash linewidth cursor pointer tandem tandem createtandem greenline var blueline new line line length stroke rgb linedash linewidth cursor pointer tandem tandem createtandem blueline var redline new line line length stroke red linedash linewidth cursor pointer tandem tandem createtandem redline i recommend factoring out a function so the lines can be created like this js var greenline createline rgb tandem createtandem greenline
| 0
|
5,382
| 8,211,044,554
|
IssuesEvent
|
2018-09-04 12:45:58
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
Document: filtering by favorite not working
|
Process bug
|
@abrahamos
open a few documents.
set one of them as a favorite.
click on filtering by favorite.
all the documents still there.

|
1.0
|
Document: filtering by favorite not working - @abrahamos
open a few documents.
set one of them as a favorite.
click on filtering by favorite.
all the documents still there.

|
process
|
document filtering by favorite not working abrahamos open a few documents set one of them as a favorite click on filtering by favorite all the documents still there
| 1
|
2,447
| 5,226,087,836
|
IssuesEvent
|
2017-01-27 20:14:34
|
AnalyticalGraphicsInc/cesium
|
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
|
closed
|
Run WebGL tests in CI
|
dev process enhancement priority
|
As discussed with @mramato offline:
- Replace all read pixels expectations with a function that can have a no-op expectation when the tests are ran with a "no WebGL" flag, e.g.,
``` javascript
expect(scene.renderForSpecs()).toEqual([0, 0, 0, 255]);
```
becomes
``` javascript
scene.expectRenderForSpecs([0, 0, 0, 255]);
```
- Replace the object returned by `getContext` with a mock object with GL functions that are no-ops, `function() {}`, except for `getExtension`, which should return mocked objects for the extensions we care about.
- Likewise, all `gl.get*` functions should be mocked to return reasonable values.
- Run the tests and fix things I forgot.
This should only take a few hours and will be more reliable than [mesa](https://github.com/AnalyticalGraphicsInc/cesium/compare/mesa).
|
1.0
|
Run WebGL tests in CI - As discussed with @mramato offline:
- Replace all read pixels expectations with a function that can have a no-op expectation when the tests are ran with a "no WebGL" flag, e.g.,
``` javascript
expect(scene.renderForSpecs()).toEqual([0, 0, 0, 255]);
```
becomes
``` javascript
scene.expectRenderForSpecs([0, 0, 0, 255]);
```
- Replace the object returned by `getContext` with a mock object with GL functions that are no-ops, `function() {}`, except for `getExtension`, which should return mocked objects for the extensions we care about.
- Likewise, all `gl.get*` functions should be mocked to return reasonable values.
- Run the tests and fix things I forgot.
This should only take a few hours and will be more reliable than [mesa](https://github.com/AnalyticalGraphicsInc/cesium/compare/mesa).
|
process
|
run webgl tests in ci as discussed with mramato offline replace all read pixels expectations with a function that can have a no op expectation when the tests are ran with a no webgl flag e g javascript expect scene renderforspecs toequal becomes javascript scene expectrenderforspecs replace the object returned by getcontext with a mock object with gl functions that are no ops function except for getextension which should return mocked objects for the extensions we care about likewise all gl get functions should be mocked to return reasonable values run the tests and fix things i forgot this should only take a few hours and will be more reliable than
| 1
|
136,522
| 11,049,379,521
|
IssuesEvent
|
2019-12-09 23:32:05
|
MangopearUK/European-Boating-Association--Theme
|
https://api.github.com/repos/MangopearUK/European-Boating-Association--Theme
|
closed
|
Test & audit: EBA subscription rate increase for budget year 2020
|
Testing: second round
|
Page URL: https://eba.eu.com/membership/secretariat-announcements/eba-subscription-rate-increase-for-budget-year-2020/
## Table of contents
- [x] **Task 1:** Perform automated audits _(10 tasks)_
- [x] **Task 2:** Manual standards & accessibility tests _(61 tasks)_
- [x] **Task 3:** Breakpoint testing _(15 tasks)_
- [x] **Task 4:** Re-run automated audits _(10 tasks)_
## 1: Perform automated audits _(10 tasks)_
### Lighthouse:
- [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_
- [x] Run "Performance" audit in lighthouse _(using incognito tab)_
- [x] Run "Best practices" audit in lighthouse _(using incognito tab)_
- [x] Run "SEO" audit in lighthouse _(using incognito tab)_
- [x] Run "PWA" audit in lighthouse _(using incognito tab)_
### Pingdom
- [x] Run full audit of the the page's performance in Pingdom
### Browser's console
- [x] Check Chrome's console for errors
### Log results of audits
- [x] Screenshot snapshot of the lighthouse audits
- [x] Upload PDF of detailed lighthouse reports
- [x] Provide a screenshot of any console errors
## 2: Manual standards & accessibility tests _(61 tasks)_
### Forms
- [x] Give all form elements permanently visible labels
- [x] Place labels above form elements
- [x] Mark invalid fields clearly and provide associated error messages
- [x] Make forms as short as possible; offer shortcuts like autocompleting the address using the postcode
- [x] Ensure all form fields have the correct requried state
- [x] Provide status and error messages as WAI-ARIA live regions
### Readability of content
- [x] Ensure page has good grammar
- [x] Ensure page content has been spell-checked
- [x] Make sure headings are in logical order
- [x] Ensure the same content is available across different devices and platforms
- [x] Begin long, multi-section documents with a table of contents
### Presentation
- [x] Make sure all content is formatted correctly
- [x] Avoid all-caps text
- [x] Make sure data tables wider than their container can be scrolled horizontally
- [x] Use the same design patterns to solve the same problems
- [x] Do not mark up subheadings/straplines with separate heading elements
### Links & buttons
#### Links
- [x] Check all links to ensure they work
- [x] Check all links to third party websites use `rel="noopener"`
- [x] Make sure the purpose of a link is clearly described: "read more" vs. "read more about accessibility"
- [x] Provide a skip link if necessary
- [x] Underline links — at least in body copy
- [x] Warn users of links that have unusual behaviors, like linking off-site, or loading a new tab (i.e. aria-label)
#### Buttons
- [x] Ensure primary calls to action are easy to recognize and reach
- [x] Provide clear, unambiguous focus styles
- [x] Ensure states (pressed, expanded, invalid, etc) are communicated to assistive software
- [x] Ensure disabled controls are not focusable
- [x] Make sure controls within hidden content are not focusable
- [x] Provide large touch "targets" for interactive elements
- [x] Make controls look like controls; give them strong perceived affordance
- [x] Use well-established, therefore recognizable, icons and symbols
### Assistive technology
- [x] Ensure content is not obscured through zooming
- [x] Support Windows high contrast mode (use images, not background images)
- [x] Provide alternative text for salient images
- [x] Make scrollable elements focusable for keyboard users
- [x] Ensure keyboard focus order is logical regarding visual layout
- [x] Match semantics to behavior for assistive technology users
- [x] Provide a default language and use lang="[ISO code]" for subsections in different languages
- [x] Inform the user when there are important changes to the application state
- [x] Do not hijack standard scrolling behavior
- [x] Do not instate "infinite scroll" by default; provide buttons to load more items
### General accessibility
- [x] Make sure text and background colors contrast sufficiently
- [x] Do not rely on color for differentiation of visual elements
- [x] Avoid images of text — text that cannot be translated, selected, or understood by assistive tech
- [x] Provide a print stylesheet
- [x] Honour requests to remove animation via the prefers-reduced-motion media query
### SEO
- [x] Ensure all pages have appropriate title
- [x] Ensure all pages have meta descriptions
- [x] Make content easier to find and improve search results with structured data [Read more](https://developers.google.com/search/docs/guides/prototype)
- [x] Check whether page should be appearing in sitemap
- [x] Make sure page has Facebook and Twitter large image previews set correctly
- [x] Check canonical links for page
- [x] Mark as cornerstone content?
### Performance
- [x] Ensure all CSS assets are minified and concatenated
- [x] Ensure all JS assets are minified and concatenated
- [x] Ensure all images are compressed
- [x] Where possible, remove redundant code
- [x] Ensure all SVG assets have been optimised
- [x] Make sure styles and scripts are not render blocking
- [x] Ensure large image assets are lazy loaded
### Other
- [x] Make sure all content belongs to a landmark element
- [x] Provide a manifest.json file for identifiable homescreen entries
## 3: Breakpoint testing _(15 tasks)_
### Desktop
- [x] Provide a full screenshot of **1920px** wide page
- [x] Provide a full screenshot of **1500px** wide page
- [x] Provide a full screenshot of **1280px** wide page
- [x] Provide a full screenshot of **1024px** wide page
### Tablet
- [x] Provide a full screenshot of **960px** wide page
- [x] Provide a full screenshot of **800px** wide page
- [x] Provide a full screenshot of **760px** wide page
- [x] Provide a full screenshot of **650px** wide page
### Mobile
- [x] Provide a full screenshot of **600px** wide page
- [x] Provide a full screenshot of **500px** wide page
- [x] Provide a full screenshot of **450px** wide page
- [x] Provide a full screenshot of **380px** wide page
- [x] Provide a full screenshot of **320px** wide page
- [x] Provide a full screenshot of **280px** wide page
- [x] Provide a full screenshot of **250px** wide page
## 4: Re-run automated audits _(10 tasks)_
### Lighthouse:
- [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_
- [x] Run "Performance" audit in lighthouse _(using incognito tab)_
- [x] Run "Best practices" audit in lighthouse _(using incognito tab)_
- [x] Run "SEO" audit in lighthouse _(using incognito tab)_
- [x] Run "PWA" audit in lighthouse _(using incognito tab)_
### Pingdom
- [x] Run full audit of the the page's performance in Pingdom
### Browser's console
- [x] Check Chrome's console for errors
### Log results of audits
- [x] Screenshot snapshot of the lighthouse audits
- [x] Upload PDF of detailed lighthouse reports
- [x] Provide a screenshot of any console errors
|
1.0
|
Test & audit: EBA subscription rate increase for budget year 2020 - Page URL: https://eba.eu.com/membership/secretariat-announcements/eba-subscription-rate-increase-for-budget-year-2020/
## Table of contents
- [x] **Task 1:** Perform automated audits _(10 tasks)_
- [x] **Task 2:** Manual standards & accessibility tests _(61 tasks)_
- [x] **Task 3:** Breakpoint testing _(15 tasks)_
- [x] **Task 4:** Re-run automated audits _(10 tasks)_
## 1: Perform automated audits _(10 tasks)_
### Lighthouse:
- [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_
- [x] Run "Performance" audit in lighthouse _(using incognito tab)_
- [x] Run "Best practices" audit in lighthouse _(using incognito tab)_
- [x] Run "SEO" audit in lighthouse _(using incognito tab)_
- [x] Run "PWA" audit in lighthouse _(using incognito tab)_
### Pingdom
- [x] Run full audit of the the page's performance in Pingdom
### Browser's console
- [x] Check Chrome's console for errors
### Log results of audits
- [x] Screenshot snapshot of the lighthouse audits
- [x] Upload PDF of detailed lighthouse reports
- [x] Provide a screenshot of any console errors
## 2: Manual standards & accessibility tests _(61 tasks)_
### Forms
- [x] Give all form elements permanently visible labels
- [x] Place labels above form elements
- [x] Mark invalid fields clearly and provide associated error messages
- [x] Make forms as short as possible; offer shortcuts like autocompleting the address using the postcode
- [x] Ensure all form fields have the correct requried state
- [x] Provide status and error messages as WAI-ARIA live regions
### Readability of content
- [x] Ensure page has good grammar
- [x] Ensure page content has been spell-checked
- [x] Make sure headings are in logical order
- [x] Ensure the same content is available across different devices and platforms
- [x] Begin long, multi-section documents with a table of contents
### Presentation
- [x] Make sure all content is formatted correctly
- [x] Avoid all-caps text
- [x] Make sure data tables wider than their container can be scrolled horizontally
- [x] Use the same design patterns to solve the same problems
- [x] Do not mark up subheadings/straplines with separate heading elements
### Links & buttons
#### Links
- [x] Check all links to ensure they work
- [x] Check all links to third party websites use `rel="noopener"`
- [x] Make sure the purpose of a link is clearly described: "read more" vs. "read more about accessibility"
- [x] Provide a skip link if necessary
- [x] Underline links — at least in body copy
- [x] Warn users of links that have unusual behaviors, like linking off-site, or loading a new tab (i.e. aria-label)
#### Buttons
- [x] Ensure primary calls to action are easy to recognize and reach
- [x] Provide clear, unambiguous focus styles
- [x] Ensure states (pressed, expanded, invalid, etc) are communicated to assistive software
- [x] Ensure disabled controls are not focusable
- [x] Make sure controls within hidden content are not focusable
- [x] Provide large touch "targets" for interactive elements
- [x] Make controls look like controls; give them strong perceived affordance
- [x] Use well-established, therefore recognizable, icons and symbols
### Assistive technology
- [x] Ensure content is not obscured through zooming
- [x] Support Windows high contrast mode (use images, not background images)
- [x] Provide alternative text for salient images
- [x] Make scrollable elements focusable for keyboard users
- [x] Ensure keyboard focus order is logical regarding visual layout
- [x] Match semantics to behavior for assistive technology users
- [x] Provide a default language and use lang="[ISO code]" for subsections in different languages
- [x] Inform the user when there are important changes to the application state
- [x] Do not hijack standard scrolling behavior
- [x] Do not instate "infinite scroll" by default; provide buttons to load more items
### General accessibility
- [x] Make sure text and background colors contrast sufficiently
- [x] Do not rely on color for differentiation of visual elements
- [x] Avoid images of text — text that cannot be translated, selected, or understood by assistive tech
- [x] Provide a print stylesheet
- [x] Honour requests to remove animation via the prefers-reduced-motion media query
### SEO
- [x] Ensure all pages have appropriate title
- [x] Ensure all pages have meta descriptions
- [x] Make content easier to find and improve search results with structured data [Read more](https://developers.google.com/search/docs/guides/prototype)
- [x] Check whether page should be appearing in sitemap
- [x] Make sure page has Facebook and Twitter large image previews set correctly
- [x] Check canonical links for page
- [x] Mark as cornerstone content?
### Performance
- [x] Ensure all CSS assets are minified and concatenated
- [x] Ensure all JS assets are minified and concatenated
- [x] Ensure all images are compressed
- [x] Where possible, remove redundant code
- [x] Ensure all SVG assets have been optimised
- [x] Make sure styles and scripts are not render blocking
- [x] Ensure large image assets are lazy loaded
### Other
- [x] Make sure all content belongs to a landmark element
- [x] Provide a manifest.json file for identifiable homescreen entries
## 3: Breakpoint testing _(15 tasks)_
### Desktop
- [x] Provide a full screenshot of **1920px** wide page
- [x] Provide a full screenshot of **1500px** wide page
- [x] Provide a full screenshot of **1280px** wide page
- [x] Provide a full screenshot of **1024px** wide page
### Tablet
- [x] Provide a full screenshot of **960px** wide page
- [x] Provide a full screenshot of **800px** wide page
- [x] Provide a full screenshot of **760px** wide page
- [x] Provide a full screenshot of **650px** wide page
### Mobile
- [x] Provide a full screenshot of **600px** wide page
- [x] Provide a full screenshot of **500px** wide page
- [x] Provide a full screenshot of **450px** wide page
- [x] Provide a full screenshot of **380px** wide page
- [x] Provide a full screenshot of **320px** wide page
- [x] Provide a full screenshot of **280px** wide page
- [x] Provide a full screenshot of **250px** wide page
## 4: Re-run automated audits _(10 tasks)_
### Lighthouse:
- [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_
- [x] Run "Performance" audit in lighthouse _(using incognito tab)_
- [x] Run "Best practices" audit in lighthouse _(using incognito tab)_
- [x] Run "SEO" audit in lighthouse _(using incognito tab)_
- [x] Run "PWA" audit in lighthouse _(using incognito tab)_
### Pingdom
- [x] Run full audit of the the page's performance in Pingdom
### Browser's console
- [x] Check Chrome's console for errors
### Log results of audits
- [x] Screenshot snapshot of the lighthouse audits
- [x] Upload PDF of detailed lighthouse reports
- [x] Provide a screenshot of any console errors
|
non_process
|
test audit eba subscription rate increase for budget year page url table of contents task perform automated audits tasks task manual standards accessibility tests tasks task breakpoint testing tasks task re run automated audits tasks perform automated audits tasks lighthouse run accessibility audit in lighthouse using incognito tab run performance audit in lighthouse using incognito tab run best practices audit in lighthouse using incognito tab run seo audit in lighthouse using incognito tab run pwa audit in lighthouse using incognito tab pingdom run full audit of the the page s performance in pingdom browser s console check chrome s console for errors log results of audits screenshot snapshot of the lighthouse audits upload pdf of detailed lighthouse reports provide a screenshot of any console errors manual standards accessibility tests tasks forms give all form elements permanently visible labels place labels above form elements mark invalid fields clearly and provide associated error messages make forms as short as possible offer shortcuts like autocompleting the address using the postcode ensure all form fields have the correct requried state provide status and error messages as wai aria live regions readability of content ensure page has good grammar ensure page content has been spell checked make sure headings are in logical order ensure the same content is available across different devices and platforms begin long multi section documents with a table of contents presentation make sure all content is formatted correctly avoid all caps text make sure data tables wider than their container can be scrolled horizontally use the same design patterns to solve the same problems do not mark up subheadings straplines with separate heading elements links buttons links check all links to ensure they work check all links to third party websites use rel noopener make sure the purpose of a link is clearly described read more vs read more about accessibility provide a skip link if necessary underline links — at least in body copy warn users of links that have unusual behaviors like linking off site or loading a new tab i e aria label buttons ensure primary calls to action are easy to recognize and reach provide clear unambiguous focus styles ensure states pressed expanded invalid etc are communicated to assistive software ensure disabled controls are not focusable make sure controls within hidden content are not focusable provide large touch targets for interactive elements make controls look like controls give them strong perceived affordance use well established therefore recognizable icons and symbols assistive technology ensure content is not obscured through zooming support windows high contrast mode use images not background images provide alternative text for salient images make scrollable elements focusable for keyboard users ensure keyboard focus order is logical regarding visual layout match semantics to behavior for assistive technology users provide a default language and use lang for subsections in different languages inform the user when there are important changes to the application state do not hijack standard scrolling behavior do not instate infinite scroll by default provide buttons to load more items general accessibility make sure text and background colors contrast sufficiently do not rely on color for differentiation of visual elements avoid images of text — text that cannot be translated selected or understood by assistive tech provide a print stylesheet honour requests to remove animation via the prefers reduced motion media query seo ensure all pages have appropriate title ensure all pages have meta descriptions make content easier to find and improve search results with structured data check whether page should be appearing in sitemap make sure page has facebook and twitter large image previews set correctly check canonical links for page mark as cornerstone content performance ensure all css assets are minified and concatenated ensure all js assets are minified and concatenated ensure all images are compressed where possible remove redundant code ensure all svg assets have been optimised make sure styles and scripts are not render blocking ensure large image assets are lazy loaded other make sure all content belongs to a landmark element provide a manifest json file for identifiable homescreen entries breakpoint testing tasks desktop provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page tablet provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page mobile provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page re run automated audits tasks lighthouse run accessibility audit in lighthouse using incognito tab run performance audit in lighthouse using incognito tab run best practices audit in lighthouse using incognito tab run seo audit in lighthouse using incognito tab run pwa audit in lighthouse using incognito tab pingdom run full audit of the the page s performance in pingdom browser s console check chrome s console for errors log results of audits screenshot snapshot of the lighthouse audits upload pdf of detailed lighthouse reports provide a screenshot of any console errors
| 0
|
37,671
| 10,057,386,823
|
IssuesEvent
|
2019-07-22 11:30:02
|
haskell/haskell-ide-engine
|
https://api.github.com/repos/haskell/haskell-ide-engine
|
closed
|
unable to build with cabal
|
build
|
as part of #1221, I need to build `hie` with cabal, which worked fine up to now, but I am getting the following error:
```
$ cabal new-build -w ghc-8.6.5
Resolving dependencies...
cabal: Could not resolve dependencies:
[__0] trying: HaRe-0.8.4.1 (user goal)
[__1] rejecting: HaRe:!test (constraint from config file, command line flag,
or user target requires opposite flag selection)
[__1] trying: HaRe:*test
[__2] trying: turtle-1.5.14 (dependency of HaRe *test)
[__3] trying: temporary-1.3 (dependency of turtle)
[__4] next goal: cabal-helper (user goal)
[__4] rejecting: cabal-helper-0.9.0.0 (conflict: temporary==1.3, cabal-helper
=> temporary<1.3 && >=1.2.1)
[__4] rejecting: cabal-helper-0.8.2.0, cabal-helper-0.8.1.2,
cabal-helper-0.8.0.2, cabal-helper-0.8.0.1, cabal-helper-0.8.0.0,
cabal-helper-0.7.3.0, cabal-helper-0.7.2.0, cabal-helper-0.7.1.0,
cabal-helper-0.6.3.1, cabal-helper-0.6.3.0, cabal-helper-0.6.2.0,
cabal-helper-0.6.1.0, cabal-helper-0.6.0.0, cabal-helper-0.5.3.0,
cabal-helper-0.5.1.0, cabal-helper-0.5.0.0, cabal-helper-0.4.0.0,
cabal-helper-0.3.9.0, cabal-helper-0.3.8.0, cabal-helper-0.3.7.0,
cabal-helper-0.3.6.0, cabal-helper-0.3.5.0, cabal-helper-0.3.4.0,
cabal-helper-0.3.3.0, cabal-helper-0.3.2.1, cabal-helper-0.3.2.0,
cabal-helper-0.3.1.0, cabal-helper-0.3.0.0, cabal-helper-0.2.0.0,
cabal-helper-0.1.0.1, cabal-helper-0.1.0.0, cabal-helper-0.8.1.1,
cabal-helper-0.8.1.0, cabal-helper-0.7.0.1, cabal-helper-0.5.2.0 (constraint
from user target requires ==0.9.0.0)
[__4] fail (backjumping, conflict set: cabal-helper, temporary)
Backjump limit reached (currently 2000, change with --max-backjumps or try to
run with --reorder-goals).
```
I have no problem building `hie` with stack.
I am on the current master (e2f8e321bbb1bff3fc5e1d2dae649a0d61a378b1). Can anyone reproduce this issue?
|
1.0
|
unable to build with cabal - as part of #1221, I need to build `hie` with cabal, which worked fine up to now, but I am getting the following error:
```
$ cabal new-build -w ghc-8.6.5
Resolving dependencies...
cabal: Could not resolve dependencies:
[__0] trying: HaRe-0.8.4.1 (user goal)
[__1] rejecting: HaRe:!test (constraint from config file, command line flag,
or user target requires opposite flag selection)
[__1] trying: HaRe:*test
[__2] trying: turtle-1.5.14 (dependency of HaRe *test)
[__3] trying: temporary-1.3 (dependency of turtle)
[__4] next goal: cabal-helper (user goal)
[__4] rejecting: cabal-helper-0.9.0.0 (conflict: temporary==1.3, cabal-helper
=> temporary<1.3 && >=1.2.1)
[__4] rejecting: cabal-helper-0.8.2.0, cabal-helper-0.8.1.2,
cabal-helper-0.8.0.2, cabal-helper-0.8.0.1, cabal-helper-0.8.0.0,
cabal-helper-0.7.3.0, cabal-helper-0.7.2.0, cabal-helper-0.7.1.0,
cabal-helper-0.6.3.1, cabal-helper-0.6.3.0, cabal-helper-0.6.2.0,
cabal-helper-0.6.1.0, cabal-helper-0.6.0.0, cabal-helper-0.5.3.0,
cabal-helper-0.5.1.0, cabal-helper-0.5.0.0, cabal-helper-0.4.0.0,
cabal-helper-0.3.9.0, cabal-helper-0.3.8.0, cabal-helper-0.3.7.0,
cabal-helper-0.3.6.0, cabal-helper-0.3.5.0, cabal-helper-0.3.4.0,
cabal-helper-0.3.3.0, cabal-helper-0.3.2.1, cabal-helper-0.3.2.0,
cabal-helper-0.3.1.0, cabal-helper-0.3.0.0, cabal-helper-0.2.0.0,
cabal-helper-0.1.0.1, cabal-helper-0.1.0.0, cabal-helper-0.8.1.1,
cabal-helper-0.8.1.0, cabal-helper-0.7.0.1, cabal-helper-0.5.2.0 (constraint
from user target requires ==0.9.0.0)
[__4] fail (backjumping, conflict set: cabal-helper, temporary)
Backjump limit reached (currently 2000, change with --max-backjumps or try to
run with --reorder-goals).
```
I have no problem building `hie` with stack.
I am on the current master (e2f8e321bbb1bff3fc5e1d2dae649a0d61a378b1). Can anyone reproduce this issue?
|
non_process
|
unable to build with cabal as part of i need to build hie with cabal which worked fine up to now but i am getting the following error cabal new build w ghc resolving dependencies cabal could not resolve dependencies trying hare user goal rejecting hare test constraint from config file command line flag or user target requires opposite flag selection trying hare test trying turtle dependency of hare test trying temporary dependency of turtle next goal cabal helper user goal rejecting cabal helper conflict temporary cabal helper temporary rejecting cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper constraint from user target requires fail backjumping conflict set cabal helper temporary backjump limit reached currently change with max backjumps or try to run with reorder goals i have no problem building hie with stack i am on the current master can anyone reproduce this issue
| 0
|
16,436
| 21,316,757,913
|
IssuesEvent
|
2022-04-16 12:16:29
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Issue with copy-to on sub-topics with locktitles
|
bug priority/medium preprocess preprocess2 stale
|
## Expected Behavior
The topic should use the navtitle defined in the map when @locktitle=yes.
Note: @copy-to and @locktitle works when used at the same level as the original topic.
## Actual Behavior
Instead topic is using the title of the first instance of the topic when topic is referenced as a sub-topic.
## Possible Solution
Fix copy-to to use the navtitle defined in the map
## Steps to Reproduce
Create ditamap with topic re-used as a sub topic
<map class="- map/map "
domains="(map bookmap) (map mapgroup-d) (map subjectScheme) (topic abbrev-d) (topic delay-d) a(props deliveryTarget) (map ditavalref-d) (map glossref-d) (topic hazard-d) (topic hi-d) (topic indexing-d) (topic markup-d) (topic pr-d) (topic relmgmt-d) (topic sw-d) (topic ui-d) (topic ut-d) (topic markup-d xml-d) (topic xnal-d) a(loc ixia_locid)"
id="axh1430149607893" title="Arron Test Map" xml:lang="en-us">
<topicref class="- map/topicref " ixia_locid="5" keys="axk1425415118966"
href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle " ixia_locid="7">Performing reading tasks</navtitle>
<linktext class="- map/linktext " ixia_locid="8">Performing reading tasks</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition,
digital dictation, or typing reports, you report on studies by way of reading tasks.
A single reading task can contain one or more studies for which you create a single
report.</shortdesc>
</topicmeta>
</topicref>
<topicref class="- map/topicref " copy-to="second-axk1425415118966.xml" ixia_locid="11"
keys="second-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle " ixia_locid="12">Arron Testing Copy-to and
Locktitle</navtitle>
<linktext class="- map/linktext " ixia_locid="13">Arron Testing Copy-to and
Locktitle</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition,
digital dictation, or typing reports, you report on studies by way of reading tasks.
A single reading task can contain one or more studies for which you create a single
report.</shortdesc>
</topicmeta>
<topicref class="- map/topicref " copy-to="third-axk1425415118966.xml" ixia_locid="14"
keys="third-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle ">Arron 2 Testing Copy-to and
Locktitle</navtitle>
<linktext class="- map/linktext " ixia_locid="16">Arron 2 Testing Copy-to and
Locktitle</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech
recognition, digital dictation, or typing reports, you report on studies by way
of reading tasks. A single reading task can contain one or more studies for
which you create a single report.</shortdesc>
</topicmeta>
</topicref>
</topicref>
<topicref class="- map/topicref " copy-to="fourth-axk1425415118966.xml" ixia_locid="14"
keys="fourth-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle ">Arron 3 Testing Copy-to and
Locktitle</navtitle>
<linktext class="- map/linktext " ixia_locid="16">Arron 3 Testing Copy-to and
Locktitle</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech
recognition, digital dictation, or typing reports, you report on studies by way
of reading tasks. A single reading task can contain one or more studies for
which you create a single report.</shortdesc>
</topicmeta>
</topicref>
</map>
<!-- Create a Gist via <https://gist.github.com/> to upload your test files. -->
<!-- Link to the Gist from the issue or attach a .zip archive of your files. -->
## Copy of the error message, log file or stack trace
None
## Environment
* DITA-OT version: 3.3.4
* Operating system and version:windows
* How did you run DITA-OT?
oxygen
* Transformation type:
HTML5
|
2.0
|
Issue with copy-to on sub-topics with locktitles - ## Expected Behavior
The topic should use the navtitle defined in the map when @locktitle=yes.
Note: @copy-to and @locktitle works when used at the same level as the original topic.
## Actual Behavior
Instead topic is using the title of the first instance of the topic when topic is referenced as a sub-topic.
## Possible Solution
Fix copy-to to use the navtitle defined in the map
## Steps to Reproduce
Create ditamap with topic re-used as a sub topic
<map class="- map/map "
domains="(map bookmap) (map mapgroup-d) (map subjectScheme) (topic abbrev-d) (topic delay-d) a(props deliveryTarget) (map ditavalref-d) (map glossref-d) (topic hazard-d) (topic hi-d) (topic indexing-d) (topic markup-d) (topic pr-d) (topic relmgmt-d) (topic sw-d) (topic ui-d) (topic ut-d) (topic markup-d xml-d) (topic xnal-d) a(loc ixia_locid)"
id="axh1430149607893" title="Arron Test Map" xml:lang="en-us">
<topicref class="- map/topicref " ixia_locid="5" keys="axk1425415118966"
href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle " ixia_locid="7">Performing reading tasks</navtitle>
<linktext class="- map/linktext " ixia_locid="8">Performing reading tasks</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition,
digital dictation, or typing reports, you report on studies by way of reading tasks.
A single reading task can contain one or more studies for which you create a single
report.</shortdesc>
</topicmeta>
</topicref>
<topicref class="- map/topicref " copy-to="second-axk1425415118966.xml" ixia_locid="11"
keys="second-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle " ixia_locid="12">Arron Testing Copy-to and
Locktitle</navtitle>
<linktext class="- map/linktext " ixia_locid="13">Arron Testing Copy-to and
Locktitle</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition,
digital dictation, or typing reports, you report on studies by way of reading tasks.
A single reading task can contain one or more studies for which you create a single
report.</shortdesc>
</topicmeta>
<topicref class="- map/topicref " copy-to="third-axk1425415118966.xml" ixia_locid="14"
keys="third-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle ">Arron 2 Testing Copy-to and
Locktitle</navtitle>
<linktext class="- map/linktext " ixia_locid="16">Arron 2 Testing Copy-to and
Locktitle</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech
recognition, digital dictation, or typing reports, you report on studies by way
of reading tasks. A single reading task can contain one or more studies for
which you create a single report.</shortdesc>
</topicmeta>
</topicref>
</topicref>
<topicref class="- map/topicref " copy-to="fourth-axk1425415118966.xml" ixia_locid="14"
keys="fourth-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task">
<topicmeta class="- map/topicmeta ">
<navtitle class="- topic/navtitle ">Arron 3 Testing Copy-to and
Locktitle</navtitle>
<linktext class="- map/linktext " ixia_locid="16">Arron 3 Testing Copy-to and
Locktitle</linktext>
<shortdesc class="- map/shortdesc ">Whether dictating with online speech
recognition, digital dictation, or typing reports, you report on studies by way
of reading tasks. A single reading task can contain one or more studies for
which you create a single report.</shortdesc>
</topicmeta>
</topicref>
</map>
<!-- Create a Gist via <https://gist.github.com/> to upload your test files. -->
<!-- Link to the Gist from the issue or attach a .zip archive of your files. -->
## Copy of the error message, log file or stack trace
None
## Environment
* DITA-OT version: 3.3.4
* Operating system and version:windows
* How did you run DITA-OT?
oxygen
* Transformation type:
HTML5
|
process
|
issue with copy to on sub topics with locktitles expected behavior the topic should use the navtitle defined in the map when locktitle yes note copy to and locktitle works when used at the same level as the original topic actual behavior instead topic is using the title of the first instance of the topic when topic is referenced as a sub topic possible solution fix copy to to use the navtitle defined in the map steps to reproduce create ditamap with topic re used as a sub topic map class map map domains map bookmap map mapgroup d map subjectscheme topic abbrev d topic delay d a props deliverytarget map ditavalref d map glossref d topic hazard d topic hi d topic indexing d topic markup d topic pr d topic relmgmt d topic sw d topic ui d topic ut d topic markup d xml d topic xnal d a loc ixia locid id title arron test map xml lang en us topicref class map topicref ixia locid keys href xml type task performing reading tasks performing reading tasks whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report topicref class map topicref copy to second xml ixia locid keys second locktitle yes href xml type task arron testing copy to and locktitle arron testing copy to and locktitle whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report topicref class map topicref copy to third xml ixia locid keys third locktitle yes href xml type task arron testing copy to and locktitle arron testing copy to and locktitle whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report topicref class map topicref copy to fourth xml ixia locid keys fourth locktitle yes href xml type task arron testing copy to and locktitle arron testing copy to and locktitle whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report copy of the error message log file or stack trace none environment dita ot version operating system and version windows how did you run dita ot oxygen transformation type
| 1
|
726,607
| 25,004,794,859
|
IssuesEvent
|
2022-11-03 10:56:39
|
sapa/performing-arts-ch-templates
|
https://api.github.com/repos/sapa/performing-arts-ch-templates
|
closed
|
New Field : rico:hasOrHadIdentifier
|
Neues Feld Priority A ToValidate Blocker BR-02 BR-03 BR-04
|
1. Creation Properties
2. Add: Form:Record Set, Record, Instantation
3. Display : Template:Record Set, Record, Instantation
## Class of **Properties** zu erstellen
- rico:hasOrHadIdentifier
-- URI : https://www.ica.org/standards/RiC/ontology#hasOrHadIdentifier
-- rdf:type : ?owl:ObjectProperty
-- rdfs:label : has or had identifier@en
- rico:isIdentifierTypeOf
-- URI : https://www.ica.org/standards/RiC/ontology#isIdentifierTypeOf
-- rdf.type : ?owl:ObjectProperty
-- rdfs:label : is Identifier Type of @en
## Class of **Entities** zu erstellen
- rico:Identifier mit properties:
-- URI : https://www.ica.org/standards/RiC/ontology#Identifier
-- rdf:type ?owl:Class
-- rdfs:label Identifier@en
- rico:IdentifierType mit proterties:
-- URI : https://www.ica.org/standards/RiC/ontology#IdentifierType
-- rdf:type ?owl:Class
-- rdfs:label : Identifier Type@en
## Vocabulary zu erstellen
- New Identifier vocabulary URI : http://vocab.performing-arts.ch/id
-- Entity: **old identifier**
--- rdf:type rico:IdentifierTyp
--- URI : http://vocab.performing-arts.ch/idold
--- rdfs:label : old identifier@en ancien identifiant@fr alt Signatur@de
-- Entity : **Unique ID**
--- rdfs:label : Unique ID@en Cote d'identification@fr Identifizierung Signatur@de
--- URI : http://vocab.performing-arts.ch/iduni
-- Entity: **technical Identifier**
--- rdf:type rico:IdentifierTyp
--- URI : http://vocab.performing-arts.ch/idtec
--- rdfs:label : Technical Identifier@en Identifiant technique@fr (à définir)@de
-- Entity: **contextual Identifier**
--- rdf:type rico:IdentifierTyp
--- URI : http://vocab.performing-arts.ch/idcon
--- rdfs:label : Contextual Identifier@en Cote de contexte@fr (à définir)@de
## Form zu modifizieren: Record Set, Record (and RecordPart), Instantation
- Hinzufügen (wie Add Daten oder Add Extent) module "Add Identifier" (add property rico:hasOrHadIdentifier and create a rico:Identifier Entity URI: http://data.performing-arts.ch/x/UUID) mit :
-- Wahl zwischen vorerstellen vocabulary von rico:IdentifierType Entities (ex: Old Identifier, Cote d'individualisation)->rico:isIdentifierTypeOf
-- string value ->rdf:value
## Template
Display the value of rico:hasOrHadIdentifier In Record Set, Record and Instantation Template under the existing label "Identifier" with the type between brackets.
## to do after
rico:identifier #166 -> #165-> #167->#168->#168
spao:legacyidentifier #170
RecordPart Order : #177
|
1.0
|
New Field : rico:hasOrHadIdentifier - 1. Creation Properties
2. Add: Form:Record Set, Record, Instantation
3. Display : Template:Record Set, Record, Instantation
## Class of **Properties** zu erstellen
- rico:hasOrHadIdentifier
-- URI : https://www.ica.org/standards/RiC/ontology#hasOrHadIdentifier
-- rdf:type : ?owl:ObjectProperty
-- rdfs:label : has or had identifier@en
- rico:isIdentifierTypeOf
-- URI : https://www.ica.org/standards/RiC/ontology#isIdentifierTypeOf
-- rdf.type : ?owl:ObjectProperty
-- rdfs:label : is Identifier Type of @en
## Class of **Entities** zu erstellen
- rico:Identifier mit properties:
-- URI : https://www.ica.org/standards/RiC/ontology#Identifier
-- rdf:type ?owl:Class
-- rdfs:label Identifier@en
- rico:IdentifierType mit proterties:
-- URI : https://www.ica.org/standards/RiC/ontology#IdentifierType
-- rdf:type ?owl:Class
-- rdfs:label : Identifier Type@en
## Vocabulary zu erstellen
- New Identifier vocabulary URI : http://vocab.performing-arts.ch/id
-- Entity: **old identifier**
--- rdf:type rico:IdentifierTyp
--- URI : http://vocab.performing-arts.ch/idold
--- rdfs:label : old identifier@en ancien identifiant@fr alt Signatur@de
-- Entity : **Unique ID**
--- rdfs:label : Unique ID@en Cote d'identification@fr Identifizierung Signatur@de
--- URI : http://vocab.performing-arts.ch/iduni
-- Entity: **technical Identifier**
--- rdf:type rico:IdentifierTyp
--- URI : http://vocab.performing-arts.ch/idtec
--- rdfs:label : Technical Identifier@en Identifiant technique@fr (à définir)@de
-- Entity: **contextual Identifier**
--- rdf:type rico:IdentifierTyp
--- URI : http://vocab.performing-arts.ch/idcon
--- rdfs:label : Contextual Identifier@en Cote de contexte@fr (à définir)@de
## Form zu modifizieren: Record Set, Record (and RecordPart), Instantation
- Hinzufügen (wie Add Daten oder Add Extent) module "Add Identifier" (add property rico:hasOrHadIdentifier and create a rico:Identifier Entity URI: http://data.performing-arts.ch/x/UUID) mit :
-- Wahl zwischen vorerstellen vocabulary von rico:IdentifierType Entities (ex: Old Identifier, Cote d'individualisation)->rico:isIdentifierTypeOf
-- string value ->rdf:value
## Template
Display the value of rico:hasOrHadIdentifier In Record Set, Record and Instantation Template under the existing label "Identifier" with the type between brackets.
## to do after
rico:identifier #166 -> #165-> #167->#168->#168
spao:legacyidentifier #170
RecordPart Order : #177
|
non_process
|
new field rico hasorhadidentifier creation properties add form record set record instantation display template record set record instantation class of properties zu erstellen rico hasorhadidentifier uri rdf type owl objectproperty rdfs label has or had identifier en rico isidentifiertypeof uri rdf type owl objectproperty rdfs label is identifier type of en class of entities zu erstellen rico identifier mit properties uri rdf type owl class rdfs label identifier en rico identifiertype mit proterties uri rdf type owl class rdfs label identifier type en vocabulary zu erstellen new identifier vocabulary uri entity old identifier rdf type rico identifiertyp uri rdfs label old identifier en ancien identifiant fr alt signatur de entity unique id rdfs label unique id en cote d identification fr identifizierung signatur de uri entity technical identifier rdf type rico identifiertyp uri rdfs label technical identifier en identifiant technique fr à définir de entity contextual identifier rdf type rico identifiertyp uri rdfs label contextual identifier en cote de contexte fr à définir de form zu modifizieren record set record and recordpart instantation hinzufügen wie add daten oder add extent module add identifier add property rico hasorhadidentifier and create a rico identifier entity uri mit wahl zwischen vorerstellen vocabulary von rico identifiertype entities ex old identifier cote d individualisation rico isidentifiertypeof string value rdf value template display the value of rico hasorhadidentifier in record set record and instantation template under the existing label identifier with the type between brackets to do after rico identifier spao legacyidentifier recordpart order
| 0
|
13,838
| 16,600,558,174
|
IssuesEvent
|
2021-06-01 18:49:09
|
Jeffail/benthos
|
https://api.github.com/repos/Jeffail/benthos
|
closed
|
Add proper argument typing to `sql` components
|
annoying bughancement inputs processors
|
Hi, I found that in the sql plugin, it converts all data to string and then exec the sql query. However, for clickhouse it will report an error. Do you think it is possible to fix this issue?
|
1.0
|
Add proper argument typing to `sql` components - Hi, I found that in the sql plugin, it converts all data to string and then exec the sql query. However, for clickhouse it will report an error. Do you think it is possible to fix this issue?
|
process
|
add proper argument typing to sql components hi i found that in the sql plugin it converts all data to string and then exec the sql query however for clickhouse it will report an error do you think it is possible to fix this issue
| 1
|
1,711
| 4,351,016,870
|
IssuesEvent
|
2016-07-31 16:20:56
|
spootTheLousy/saguaro
|
https://api.github.com/repos/spootTheLousy/saguaro
|
opened
|
Check images for hidden archives
|
Image processing Security
|
Currently, saguaro does no sanitation or checks in regards to images that have hidden embedded content. More often than not, it's a method of transmitting illegal content, so it is definitely within the scope of the repo to check images for this.
Fortunately, PHP 5.2 has two classes that ship with it for handling `.rar` and `.zip` files, should be as simple as checking if an image is extractable, and if it is, rejecting it.
|
1.0
|
Check images for hidden archives - Currently, saguaro does no sanitation or checks in regards to images that have hidden embedded content. More often than not, it's a method of transmitting illegal content, so it is definitely within the scope of the repo to check images for this.
Fortunately, PHP 5.2 has two classes that ship with it for handling `.rar` and `.zip` files, should be as simple as checking if an image is extractable, and if it is, rejecting it.
|
process
|
check images for hidden archives currently saguaro does no sanitation or checks in regards to images that have hidden embedded content more often than not it s a method of transmitting illegal content so it is definitely within the scope of the repo to check images for this fortunately php has two classes that ship with it for handling rar and zip files should be as simple as checking if an image is extractable and if it is rejecting it
| 1
|
16,620
| 9,853,207,050
|
IssuesEvent
|
2019-06-19 14:21:35
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
Object Level Security
|
Team:Security enhancement
|
## Object Level Security
### ACL
To enable OLS, an ACL with the following format will be attached to all securable objects:
```
{
"owner": 123456789,
"read": {
"users": [ { id: 123456789, can_share: true } ],
"roles": [ { id: "role_one", can_share: false } ]
},
"write": {
"users": [ { id: 123456789, can_share: true } ],
"roles": [ { id: "role_one", can_share: false } ]
}
}
```
The users and roles above reference Elasticsearch users and roles. In addition to the Elasticsearch roles, we will utilize a role of `*` to denote all authenticated users.
If the user has **write** they will implicitly be granted **read**. When a user or role is assigned **read** or **write** they will be able to specify whether this user or role will be able to share the securable object with others. If the user has **read** and they can share the object, they will only be able to add other users and roles to **read**. If the user has **write** and they can share the object, they will be able to add other users and roles to **write** and **read**.
### Implicit read permissions
When a user has **read** access to a Dashboard, they will implicitly be granted **read** access to all related Visualizations and Saved Searches. The same logic will apply once Index Patterns themselves are made securable and if a user has **read** access to a Visualization or Saved Search, they will be implicitly granted **read** access to the Index Pattern.
This simplifies the access model and allows users to assign access to the object that they intuitively wish to share without having to concern themselves with the graph of related objects. It also simplifies the technical implementation so we don’t have to explicitly assign access to the related objects and then determine if/when it should be removed when a parent object’s ACL is modified.
When a user is implicitly granted **read** access to a Visualization or Saved Search, it won’t show up in the user’s list of Visualizations or Saved Searches, it will only be accessible in the Dashboard UI/API. This is similar to how we’ll implement it technically, we’ll allow users to gain access to the related objects via the Dashboard, which will implicitly be granting them **read** access.
### Summary
Phase 4 will make Saved Searches, Dashboards, Visualizations, Index Patterns and other Kibana applications (Machine Learning, Graph, Timelion) saved objects securable based on the previously described ACL.
When an object has no owner, it emulates the way that Kibana currently functions without OLS where all authenticated Kibana users have full permissions. This is purely to support migrations from older versions of Kibana that didn’t have OLS, or users that were running Kibana without security and then enabling security with OLS. An additional “Claim unowned object” privilege will be added to the kibana_user role, and the user will have to have this privilege to claim these unowned objects.
The introduction of owned Index Patterns necessitates the addition of per-user Kibana Advanced Settings, as the default index pattern is defined here. An additional section will be added to the Advanced Settings page to allow a user to override any advanced setting, the same capability will be added to the index management page.
When a securable object has no owner, they will see a dialog similar to the following allowing them to make themselves the owner:

A securable object with no owner will be represented by the non-existence of an ACL.
When a securable object has an owner, they will see a dialog similar to the following allowing them to transfer ownership and define which users and roles can read/write the object:

System administrators will always be able to transfer ownership amd modify the ACL of a securable object, incase a user erroneously claims ownership of an owned object.
All users that have a role granting them a Kibana custom privilege for the specific Kibana instance will be listed, and all roles that have a Kibana custom privilege for the Kibana specific instance will be listed as well.
It should be noted that for Kibana to be able to fully enumerate users, we will have to introduce the concept of user profiles in Kibana (that could potentially power the user specific settings) or have Elasticsearch create users for non-native realms. Currently, Elasticsearch is unable to enumerate all users for SAML/LDAP/etc. realms as these are powered by role mappings.
The list of Saved Searches, Dashboards, Visualizations and Index Patterns will have an owner column added, similar to the following:

From this phase forward, all new securable objects will be owned by the creator and they will have to share them with others. This same logic applies to objects that are imported. They will be owned by the user importing them, and can then be shared.
Additional Kibana applications (Graph, Timelion) will be modified to support a similar mechanism of claiming/transferring ownership, and listing the current owner.
In the future, there’s potential for the Kibana admin to be able to define default permissions for different users, or to use RBAC to limit users being able to create private or public securable objects. However, this level of control will not be introduced in this phase, as it might not be needed and it increases the complexity and implementation time.
|
True
|
Object Level Security - ## Object Level Security
### ACL
To enable OLS, an ACL with the following format will be attached to all securable objects:
```
{
"owner": 123456789,
"read": {
"users": [ { id: 123456789, can_share: true } ],
"roles": [ { id: "role_one", can_share: false } ]
},
"write": {
"users": [ { id: 123456789, can_share: true } ],
"roles": [ { id: "role_one", can_share: false } ]
}
}
```
The users and roles above reference Elasticsearch users and roles. In addition to the Elasticsearch roles, we will utilize a role of `*` to denote all authenticated users.
If the user has **write** they will implicitly be granted **read**. When a user or role is assigned **read** or **write** they will be able to specify whether this user or role will be able to share the securable object with others. If the user has **read** and they can share the object, they will only be able to add other users and roles to **read**. If the user has **write** and they can share the object, they will be able to add other users and roles to **write** and **read**.
### Implicit read permissions
When a user has **read** access to a Dashboard, they will implicitly be granted **read** access to all related Visualizations and Saved Searches. The same logic will apply once Index Patterns themselves are made securable and if a user has **read** access to a Visualization or Saved Search, they will be implicitly granted **read** access to the Index Pattern.
This simplifies the access model and allows users to assign access to the object that they intuitively wish to share without having to concern themselves with the graph of related objects. It also simplifies the technical implementation so we don’t have to explicitly assign access to the related objects and then determine if/when it should be removed when a parent object’s ACL is modified.
When a user is implicitly granted **read** access to a Visualization or Saved Search, it won’t show up in the user’s list of Visualizations or Saved Searches, it will only be accessible in the Dashboard UI/API. This is similar to how we’ll implement it technically, we’ll allow users to gain access to the related objects via the Dashboard, which will implicitly be granting them **read** access.
### Summary
Phase 4 will make Saved Searches, Dashboards, Visualizations, Index Patterns and other Kibana applications (Machine Learning, Graph, Timelion) saved objects securable based on the previously described ACL.
When an object has no owner, it emulates the way that Kibana currently functions without OLS where all authenticated Kibana users have full permissions. This is purely to support migrations from older versions of Kibana that didn’t have OLS, or users that were running Kibana without security and then enabling security with OLS. An additional “Claim unowned object” privilege will be added to the kibana_user role, and the user will have to have this privilege to claim these unowned objects.
The introduction of owned Index Patterns necessitates the addition of per-user Kibana Advanced Settings, as the default index pattern is defined here. An additional section will be added to the Advanced Settings page to allow a user to override any advanced setting, the same capability will be added to the index management page.
When a securable object has no owner, they will see a dialog similar to the following allowing them to make themselves the owner:

A securable object with no owner will be represented by the non-existence of an ACL.
When a securable object has an owner, they will see a dialog similar to the following allowing them to transfer ownership and define which users and roles can read/write the object:

System administrators will always be able to transfer ownership amd modify the ACL of a securable object, incase a user erroneously claims ownership of an owned object.
All users that have a role granting them a Kibana custom privilege for the specific Kibana instance will be listed, and all roles that have a Kibana custom privilege for the Kibana specific instance will be listed as well.
It should be noted that for Kibana to be able to fully enumerate users, we will have to introduce the concept of user profiles in Kibana (that could potentially power the user specific settings) or have Elasticsearch create users for non-native realms. Currently, Elasticsearch is unable to enumerate all users for SAML/LDAP/etc. realms as these are powered by role mappings.
The list of Saved Searches, Dashboards, Visualizations and Index Patterns will have an owner column added, similar to the following:

From this phase forward, all new securable objects will be owned by the creator and they will have to share them with others. This same logic applies to objects that are imported. They will be owned by the user importing them, and can then be shared.
Additional Kibana applications (Graph, Timelion) will be modified to support a similar mechanism of claiming/transferring ownership, and listing the current owner.
In the future, there’s potential for the Kibana admin to be able to define default permissions for different users, or to use RBAC to limit users being able to create private or public securable objects. However, this level of control will not be introduced in this phase, as it might not be needed and it increases the complexity and implementation time.
|
non_process
|
object level security object level security acl to enable ols an acl with the following format will be attached to all securable objects owner read users roles write users roles the users and roles above reference elasticsearch users and roles in addition to the elasticsearch roles we will utilize a role of to denote all authenticated users if the user has write they will implicitly be granted read when a user or role is assigned read or write they will be able to specify whether this user or role will be able to share the securable object with others if the user has read and they can share the object they will only be able to add other users and roles to read if the user has write and they can share the object they will be able to add other users and roles to write and read implicit read permissions when a user has read access to a dashboard they will implicitly be granted read access to all related visualizations and saved searches the same logic will apply once index patterns themselves are made securable and if a user has read access to a visualization or saved search they will be implicitly granted read access to the index pattern this simplifies the access model and allows users to assign access to the object that they intuitively wish to share without having to concern themselves with the graph of related objects it also simplifies the technical implementation so we don’t have to explicitly assign access to the related objects and then determine if when it should be removed when a parent object’s acl is modified when a user is implicitly granted read access to a visualization or saved search it won’t show up in the user’s list of visualizations or saved searches it will only be accessible in the dashboard ui api this is similar to how we’ll implement it technically we’ll allow users to gain access to the related objects via the dashboard which will implicitly be granting them read access summary phase will make saved searches dashboards visualizations index patterns and other kibana applications machine learning graph timelion saved objects securable based on the previously described acl when an object has no owner it emulates the way that kibana currently functions without ols where all authenticated kibana users have full permissions this is purely to support migrations from older versions of kibana that didn’t have ols or users that were running kibana without security and then enabling security with ols an additional “claim unowned object” privilege will be added to the kibana user role and the user will have to have this privilege to claim these unowned objects the introduction of owned index patterns necessitates the addition of per user kibana advanced settings as the default index pattern is defined here an additional section will be added to the advanced settings page to allow a user to override any advanced setting the same capability will be added to the index management page when a securable object has no owner they will see a dialog similar to the following allowing them to make themselves the owner a securable object with no owner will be represented by the non existence of an acl when a securable object has an owner they will see a dialog similar to the following allowing them to transfer ownership and define which users and roles can read write the object system administrators will always be able to transfer ownership amd modify the acl of a securable object incase a user erroneously claims ownership of an owned object all users that have a role granting them a kibana custom privilege for the specific kibana instance will be listed and all roles that have a kibana custom privilege for the kibana specific instance will be listed as well it should be noted that for kibana to be able to fully enumerate users we will have to introduce the concept of user profiles in kibana that could potentially power the user specific settings or have elasticsearch create users for non native realms currently elasticsearch is unable to enumerate all users for saml ldap etc realms as these are powered by role mappings the list of saved searches dashboards visualizations and index patterns will have an owner column added similar to the following from this phase forward all new securable objects will be owned by the creator and they will have to share them with others this same logic applies to objects that are imported they will be owned by the user importing them and can then be shared additional kibana applications graph timelion will be modified to support a similar mechanism of claiming transferring ownership and listing the current owner in the future there’s potential for the kibana admin to be able to define default permissions for different users or to use rbac to limit users being able to create private or public securable objects however this level of control will not be introduced in this phase as it might not be needed and it increases the complexity and implementation time
| 0
|
6,466
| 9,546,611,790
|
IssuesEvent
|
2019-05-01 20:27:31
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Department of State Drawers
|
Apply Process Requirements Ready State Dept.
|
Who: Student
What: ability to edit or delete information in Open Opps
Why: As a student applying for an internship I would like the ability to edit and delete information from my application
A/C
- The drawers throughout the application will follow the design system https://usajobs.github.io/design-system/drawer/
- Each drawer will for References, Experience and Languages will have an edit button that will allow a user to edit the information in Open Opportunities only
- Each drawer for References, Experience and Languages will have a delete button that will allow the user to delete the information in Open Opportunities only
|
1.0
|
Department of State Drawers - Who: Student
What: ability to edit or delete information in Open Opps
Why: As a student applying for an internship I would like the ability to edit and delete information from my application
A/C
- The drawers throughout the application will follow the design system https://usajobs.github.io/design-system/drawer/
- Each drawer will for References, Experience and Languages will have an edit button that will allow a user to edit the information in Open Opportunities only
- Each drawer for References, Experience and Languages will have a delete button that will allow the user to delete the information in Open Opportunities only
|
process
|
department of state drawers who student what ability to edit or delete information in open opps why as a student applying for an internship i would like the ability to edit and delete information from my application a c the drawers throughout the application will follow the design system each drawer will for references experience and languages will have an edit button that will allow a user to edit the information in open opportunities only each drawer for references experience and languages will have a delete button that will allow the user to delete the information in open opportunities only
| 1
|
121,162
| 10,152,085,424
|
IssuesEvent
|
2019-08-05 22:15:47
|
MadalinNitu/Bendis
|
https://api.github.com/repos/MadalinNitu/Bendis
|
closed
|
Update Profile models for allow guest accounts.
|
back-end medium testing
|
Update Profile models for allow guest accounts.
|
1.0
|
Update Profile models for allow guest accounts. - Update Profile models for allow guest accounts.
|
non_process
|
update profile models for allow guest accounts update profile models for allow guest accounts
| 0
|
99,720
| 4,063,715,432
|
IssuesEvent
|
2016-05-26 01:24:22
|
OregonCore/OregonCore
|
https://api.github.com/repos/OregonCore/OregonCore
|
closed
|
Problem with visibility
|
Category: Entities Category: Miscellaneous Priority: High Type: Bug
|
When you delete a creature, it does not dissapear until you move.
May be related to #1199
|
1.0
|
Problem with visibility - When you delete a creature, it does not dissapear until you move.
May be related to #1199
|
non_process
|
problem with visibility when you delete a creature it does not dissapear until you move may be related to
| 0
|
19,776
| 26,156,392,637
|
IssuesEvent
|
2022-12-30 22:48:19
|
kitspace/kitspace-v2
|
https://api.github.com/repos/kitspace/kitspace-v2
|
opened
|
Docker volumes size increases significantly after importing all projects
|
processor
|
I have noticed that the deployment to my staging server has been starting to fail frequently due to disk space.
Checking the disk usage I have found that the processor and gitea volumes are taking up ~**25GB** which causes image pull operations to fail.
We can either:
1. upgrade the server,
2. delete the repos from the processor in a clean-up job.
I prefer the second option.
##
```console
$ docker system df -v
Images space usage:
REPOSITORY TAG IMAGE ID CREATED SIZE SHARED SIZE UNIQUE SIZE CONTAINERS
Containers space usage:
CONTAINER ID IMAGE COMMAND LOCAL VOLUMES SIZE CREATED STATUS NAMES
Local Volumes space usage:
VOLUME NAME LINKS SIZE
kitspace-abdo-dev_gitea-data 0 8.128GB <---------
fb020e97396eef361bf914971fca183126044a8554e38dd1858e223cd8428716 0 520.1kB
kitspace-abdo-dev_postgres-data 0 76.63MB
kitspace-abdo-dev_processor-data 0 16.84GB <---------
kitspace-abdo-dev_meilisearch-data 0 71.21MB
9cb3800976bc29e7db7bb37e8f98f70322d695d0fa2994f9efdf4c1cce18e328 0 21.27kB
```
|
1.0
|
Docker volumes size increases significantly after importing all projects - I have noticed that the deployment to my staging server has been starting to fail frequently due to disk space.
Checking the disk usage I have found that the processor and gitea volumes are taking up ~**25GB** which causes image pull operations to fail.
We can either:
1. upgrade the server,
2. delete the repos from the processor in a clean-up job.
I prefer the second option.
##
```console
$ docker system df -v
Images space usage:
REPOSITORY TAG IMAGE ID CREATED SIZE SHARED SIZE UNIQUE SIZE CONTAINERS
Containers space usage:
CONTAINER ID IMAGE COMMAND LOCAL VOLUMES SIZE CREATED STATUS NAMES
Local Volumes space usage:
VOLUME NAME LINKS SIZE
kitspace-abdo-dev_gitea-data 0 8.128GB <---------
fb020e97396eef361bf914971fca183126044a8554e38dd1858e223cd8428716 0 520.1kB
kitspace-abdo-dev_postgres-data 0 76.63MB
kitspace-abdo-dev_processor-data 0 16.84GB <---------
kitspace-abdo-dev_meilisearch-data 0 71.21MB
9cb3800976bc29e7db7bb37e8f98f70322d695d0fa2994f9efdf4c1cce18e328 0 21.27kB
```
|
process
|
docker volumes size increases significantly after importing all projects i have noticed that the deployment to my staging server has been starting to fail frequently due to disk space checking the disk usage i have found that the processor and gitea volumes are taking up which causes image pull operations to fail we can either upgrade the server delete the repos from the processor in a clean up job i prefer the second option console docker system df v images space usage repository tag image id created size shared size unique size containers containers space usage container id image command local volumes size created status names local volumes space usage volume name links size kitspace abdo dev gitea data kitspace abdo dev postgres data kitspace abdo dev processor data kitspace abdo dev meilisearch data
| 1
|
15,557
| 19,703,503,472
|
IssuesEvent
|
2022-01-12 19:08:00
|
googleapis/java-securitycenter-settings
|
https://api.github.com/repos/googleapis/java-securitycenter-settings
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'securitycenter-settings' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'securitycenter-settings' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname securitycenter settings invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
688,526
| 23,586,433,079
|
IssuesEvent
|
2022-08-23 12:02:17
|
mm-ninja-turtles/turtle-express
|
https://api.github.com/repos/mm-ninja-turtles/turtle-express
|
closed
|
[bug]: in request validation
|
type: bug priority: high task: todo
|
I tested this kind of example with request body
```
const users = router.path('/users')
users.handler({
id: 'CreateNewUser',
method: 'post',
request: {
body: z.object({
name: z.string().min(1, { message: 'name is required.' }),
email: z.string().email({ message: 'wrong email format.' }),
}),
},
response: {
200: z.object({
name: z.string(),
email: z.string().email(),
}),
},
async resolver({ ctx }) {
const { name, email } = ctx.body
const user = await db.users.create({ name, email })
return {
200: user,
}
},
})
'''
|
1.0
|
[bug]: in request validation - I tested this kind of example with request body
```
const users = router.path('/users')
users.handler({
id: 'CreateNewUser',
method: 'post',
request: {
body: z.object({
name: z.string().min(1, { message: 'name is required.' }),
email: z.string().email({ message: 'wrong email format.' }),
}),
},
response: {
200: z.object({
name: z.string(),
email: z.string().email(),
}),
},
async resolver({ ctx }) {
const { name, email } = ctx.body
const user = await db.users.create({ name, email })
return {
200: user,
}
},
})
'''
|
non_process
|
in request validation i tested this kind of example with request body const users router path users users handler id createnewuser method post request body z object name z string min message name is required email z string email message wrong email format response z object name z string email z string email async resolver ctx const name email ctx body const user await db users create name email return user
| 0
|
27,197
| 6,819,828,528
|
IssuesEvent
|
2017-11-07 11:39:28
|
kimushu/rubic-vscode
|
https://api.github.com/repos/kimushu/rubic-vscode
|
closed
|
VSCode is no longer supporting specifiying a commandId in the "initialConfigurations" contribution
|
vscode
|
Hi,
We noticed your extension in `package.json` is using a `"initialConfigurations"` contribution in a deprecated way by specifiying a command id.
We wanted to let you know that we plan to delete support for this specific use of "initialConfigurations" soon.
More about this deprecation can be found [here](https://github.com/Microsoft/vscode/issues/33791) and in our release notes.
A good example on how to use the `DebugConfigurarationProvider` which is the new way of doing this can be found [here](https://github.com/Microsoft/vscode-node-debug/blob/d61ac00e16069fe63a5a5b6c65a4040de013e7f2/src/node/extension/configurationProvider.ts#L18)
Also note other deprecations which are specified in our release notes (which can be found on our site).
Kind regards,
Isidor from the VSCode team
|
1.0
|
VSCode is no longer supporting specifiying a commandId in the "initialConfigurations" contribution - Hi,
We noticed your extension in `package.json` is using a `"initialConfigurations"` contribution in a deprecated way by specifiying a command id.
We wanted to let you know that we plan to delete support for this specific use of "initialConfigurations" soon.
More about this deprecation can be found [here](https://github.com/Microsoft/vscode/issues/33791) and in our release notes.
A good example on how to use the `DebugConfigurarationProvider` which is the new way of doing this can be found [here](https://github.com/Microsoft/vscode-node-debug/blob/d61ac00e16069fe63a5a5b6c65a4040de013e7f2/src/node/extension/configurationProvider.ts#L18)
Also note other deprecations which are specified in our release notes (which can be found on our site).
Kind regards,
Isidor from the VSCode team
|
non_process
|
vscode is no longer supporting specifiying a commandid in the initialconfigurations contribution hi we noticed your extension in package json is using a initialconfigurations contribution in a deprecated way by specifiying a command id we wanted to let you know that we plan to delete support for this specific use of initialconfigurations soon more about this deprecation can be found and in our release notes a good example on how to use the debugconfigurarationprovider which is the new way of doing this can be found also note other deprecations which are specified in our release notes which can be found on our site kind regards isidor from the vscode team
| 0
|
15,992
| 20,188,203,910
|
IssuesEvent
|
2022-02-11 01:17:45
|
savitamittalmsft/WAS-SEC-TEST
|
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
|
opened
|
Standardize on modern authentication protocols
|
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Authentication and authorization
|
<a href="https://docs.microsoft.com/azure/architecture/framework/security/design-identity-authentication#use-modern-password-protection">Standardize on modern authentication protocols</a>
<p><b>Why Consider This?</b></p>
Modern authentication protocols support strong controls such as MFA and should be used instead of legacy authentication methods. Use of legacy method increases risk of credential exposure.
<p><b>Context</b></p>
<p><span>Modern Authentication is a method of identity management that offers more secure user authentication and authorization. It's available for Office 365 hybrid deployments of Skype for Business server on-premises and Exchange server on-premises, as well as, split-domain Skype for Business hybrids. This article links to related docs about prerequisites, setup/disabling modern authentication, and to some of the related client (ex. Outlook and Skype clients) information.</span></p><p><span>Modern authentication is an umbrella term for a combination of authentication and authorization methods between a client (for example, your laptop or your phone) and a server, as well as some security measures that rely on access policies that you may already be familiar with. It includes:</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span><b>Authentication methods</b></span><span>: Multi-factor authentication (MFA); smart card authentication; client certificate-based authentication</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span><b>Authorization methods</b></span><span>: Microsoft's implementation of Open Authorization (OAuth)</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span><b>Conditional access policies</b></span><span>: Mobile Application Management (MAM) and Azure Active Directory (Azure AD) Conditional Access</span></li></ul>
<p><b>Suggested Actions</b></p>
<p><span>Review workloads that do not leverage modern authentication protocols and convert where possible. In addition, standarize using modern authentication protocols for all future workloads.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/microsoft-365/enterprise/hybrid-modern-auth-overview?view=o365-worldwide#BKMK_WhatisModAuth" target="_blank"><span>What is modern authentication?</span></a><span /></p>
|
1.0
|
Standardize on modern authentication protocols - <a href="https://docs.microsoft.com/azure/architecture/framework/security/design-identity-authentication#use-modern-password-protection">Standardize on modern authentication protocols</a>
<p><b>Why Consider This?</b></p>
Modern authentication protocols support strong controls such as MFA and should be used instead of legacy authentication methods. Use of legacy method increases risk of credential exposure.
<p><b>Context</b></p>
<p><span>Modern Authentication is a method of identity management that offers more secure user authentication and authorization. It's available for Office 365 hybrid deployments of Skype for Business server on-premises and Exchange server on-premises, as well as, split-domain Skype for Business hybrids. This article links to related docs about prerequisites, setup/disabling modern authentication, and to some of the related client (ex. Outlook and Skype clients) information.</span></p><p><span>Modern authentication is an umbrella term for a combination of authentication and authorization methods between a client (for example, your laptop or your phone) and a server, as well as some security measures that rely on access policies that you may already be familiar with. It includes:</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span><b>Authentication methods</b></span><span>: Multi-factor authentication (MFA); smart card authentication; client certificate-based authentication</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span><b>Authorization methods</b></span><span>: Microsoft's implementation of Open Authorization (OAuth)</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span><b>Conditional access policies</b></span><span>: Mobile Application Management (MAM) and Azure Active Directory (Azure AD) Conditional Access</span></li></ul>
<p><b>Suggested Actions</b></p>
<p><span>Review workloads that do not leverage modern authentication protocols and convert where possible. In addition, standarize using modern authentication protocols for all future workloads.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/microsoft-365/enterprise/hybrid-modern-auth-overview?view=o365-worldwide#BKMK_WhatisModAuth" target="_blank"><span>What is modern authentication?</span></a><span /></p>
|
process
|
standardize on modern authentication protocols why consider this modern authentication protocols support strong controls such as mfa and should be used instead of legacy authentication methods use of legacy method increases risk of credential exposure context modern authentication is a method of identity management that offers more secure user authentication and authorization it s available for office hybrid deployments of skype for business server on premises and exchange server on premises as well as split domain skype for business hybrids this article links to related docs about prerequisites setup disabling modern authentication and to some of the related client ex outlook and skype clients information modern authentication is an umbrella term for a combination of authentication and authorization methods between a client for example your laptop or your phone and a server as well as some security measures that rely on access policies that you may already be familiar with it includes authentication methods multi factor authentication mfa smart card authentication client certificate based authentication authorization methods microsoft s implementation of open authorization oauth conditional access policies mobile application management mam and azure active directory azure ad conditional access suggested actions review workloads that do not leverage modern authentication protocols and convert where possible in addition standarize using modern authentication protocols for all future workloads learn more what is modern authentication
| 1
|
2,604
| 5,357,060,864
|
IssuesEvent
|
2017-02-20 17:14:37
|
PagerNation/PagerNation
|
https://api.github.com/repos/PagerNation/PagerNation
|
closed
|
Read UI notes & adjust
|
process
|
## Backlog
### Problem
> Give a description of the card. What is the problem you want to solve?
### Why?
> What will completion of this card gain us?
> Link back to a Use Case or Functional Requirement that this will solve
## Elaboration
### How do we solve this?
> Get as specific as possible, talk about specific files, endpoints, etc. Anyone should be able to pick up the card and complete it in basically the same way as anyone else
### When do we know this is done?
> Outline acceptance criteria
### Tests!?!!?!
> Describe the tests that need to be implemented. If it's a complicated feature, get descriptive. Otherwise it's fine to be less verbose.
|
1.0
|
Read UI notes & adjust - ## Backlog
### Problem
> Give a description of the card. What is the problem you want to solve?
### Why?
> What will completion of this card gain us?
> Link back to a Use Case or Functional Requirement that this will solve
## Elaboration
### How do we solve this?
> Get as specific as possible, talk about specific files, endpoints, etc. Anyone should be able to pick up the card and complete it in basically the same way as anyone else
### When do we know this is done?
> Outline acceptance criteria
### Tests!?!!?!
> Describe the tests that need to be implemented. If it's a complicated feature, get descriptive. Otherwise it's fine to be less verbose.
|
process
|
read ui notes adjust backlog problem give a description of the card what is the problem you want to solve why what will completion of this card gain us link back to a use case or functional requirement that this will solve elaboration how do we solve this get as specific as possible talk about specific files endpoints etc anyone should be able to pick up the card and complete it in basically the same way as anyone else when do we know this is done outline acceptance criteria tests describe the tests that need to be implemented if it s a complicated feature get descriptive otherwise it s fine to be less verbose
| 1
|
10,746
| 13,540,496,793
|
IssuesEvent
|
2020-09-16 14:44:07
|
pystatgen/sgkit
|
https://api.github.com/repos/pystatgen/sgkit
|
opened
|
Add badges to main page/readme
|
process + tools
|
We should consider adding badges for:
* master build
* master doc publish
* future validation suite
|
1.0
|
Add badges to main page/readme - We should consider adding badges for:
* master build
* master doc publish
* future validation suite
|
process
|
add badges to main page readme we should consider adding badges for master build master doc publish future validation suite
| 1
|
144,000
| 11,590,804,873
|
IssuesEvent
|
2020-02-24 07:55:13
|
INTO-CPS-Association/pyfmu
|
https://api.github.com/repos/INTO-CPS-Association/pyfmu
|
opened
|
functional testing of example projects
|
testing
|
Currently, the test suite only tests the functionality of the exported example projects by checking if the simulation is successful.
A good addition would be to add tests that trace of the simulation to ensure they behave correctly.
|
1.0
|
functional testing of example projects - Currently, the test suite only tests the functionality of the exported example projects by checking if the simulation is successful.
A good addition would be to add tests that trace of the simulation to ensure they behave correctly.
|
non_process
|
functional testing of example projects currently the test suite only tests the functionality of the exported example projects by checking if the simulation is successful a good addition would be to add tests that trace of the simulation to ensure they behave correctly
| 0
|
264,703
| 23,134,419,964
|
IssuesEvent
|
2022-07-28 13:16:16
|
etcd-io/etcd
|
https://api.github.com/repos/etcd-io/etcd
|
closed
|
FLAKE: TestKVDelete
|
Help Wanted area/testing/flake
|
https://github.com/etcd-io/etcd/runs/7499725943?check_suite_focus=true
Interesting log:
```
/home/runner/work/etcd/etcd/bin/etcd-25280: {"level":"error","ts":"2022-07-25T12:53:51.450Z","caller":"embed/etcd.go:778","msg":"setting up serving from embedded etcd failed.","error":"accept tcp 127.0.0.1:20000: use of closed network connection","stacktrace":"go.etcd.io/etcd/server/v3/embed.(*Etcd).errHandler\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:778\ngo.etcd.io/etcd/server/v3/embed.(*Etcd).serveClients.func1\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:735"}
```
|
1.0
|
FLAKE: TestKVDelete - https://github.com/etcd-io/etcd/runs/7499725943?check_suite_focus=true
Interesting log:
```
/home/runner/work/etcd/etcd/bin/etcd-25280: {"level":"error","ts":"2022-07-25T12:53:51.450Z","caller":"embed/etcd.go:778","msg":"setting up serving from embedded etcd failed.","error":"accept tcp 127.0.0.1:20000: use of closed network connection","stacktrace":"go.etcd.io/etcd/server/v3/embed.(*Etcd).errHandler\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:778\ngo.etcd.io/etcd/server/v3/embed.(*Etcd).serveClients.func1\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:735"}
```
|
non_process
|
flake testkvdelete interesting log home runner work etcd etcd bin etcd level error ts caller embed etcd go msg setting up serving from embedded etcd failed error accept tcp use of closed network connection stacktrace go etcd io etcd server embed etcd errhandler n t home runner work etcd etcd server embed etcd go ngo etcd io etcd server embed etcd serveclients n t home runner work etcd etcd server embed etcd go
| 0
|
350,989
| 31,932,708,394
|
IssuesEvent
|
2023-09-19 08:30:30
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix tensor.test_tensorflow_tensor_shape
|
TensorFlow Frontend Sub Task Failing Test
|
| | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6201081410/job/16837044459"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix tensor.test_tensorflow_tensor_shape - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6201081410/job/16837044459"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
|
non_process
|
fix tensor test tensorflow tensor shape numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
| 0
|
16,366
| 21,053,352,179
|
IssuesEvent
|
2022-03-31 22:53:16
|
googleapis/google-auth-library-ruby
|
https://api.github.com/repos/googleapis/google-auth-library-ruby
|
closed
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* could not parse .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* could not parse .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 could not parse repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
181,382
| 6,659,216,936
|
IssuesEvent
|
2017-10-01 08:09:12
|
k0shk0sh/FastHub
|
https://api.github.com/repos/k0shk0sh/FastHub
|
closed
|
Annoying and blocking "Too many changes to show"
|
Priority: Critical Status: Accepted Type: Enhancement
|
**FastHub Version: 4.3.0**
**Android Version: 7.0 (SDK: 24)**
**Device Information:**
- **Manufacturer:** samsung
- **Brand:** samsung
- **Model:** Galaxy Tab S2 8.0
---
It seems that you added a threshold in a new version and don't show a diff in PR if file contains many changes. This is a horrible decision that significantly decrease the usability. First of all this threshold seems to be very small as most of changes are not being displayed more. Another point is that PR review feature is affected. You sell a PRO feature to perform a code review via app, however how could it be possible to left a comment if diff isn't displayed? Now I'm able to left comments to a few files in only, while for rest I should use the browser.
I'm very disappointed by this bug :pensive: For me it's about the current approach in general - in each version you introduce some new "features", that later create the usability issues (minor or major) :cry: It's a bit annoying as I feel like a beta tester.. Of course, you quickly fix them, but for couple of weeks we should live with them :( Could you please consider adding an advanced configuration for such things, so we could quickly disable/tune new features if there issues with them? For instance, it would be enough to tune the threshold value to work around the current issue, rather than be unable to perform a code review for a few weeks..
Thank you for your time and work on this project!
|
1.0
|
Annoying and blocking "Too many changes to show" - **FastHub Version: 4.3.0**
**Android Version: 7.0 (SDK: 24)**
**Device Information:**
- **Manufacturer:** samsung
- **Brand:** samsung
- **Model:** Galaxy Tab S2 8.0
---
It seems that you added a threshold in a new version and don't show a diff in PR if file contains many changes. This is a horrible decision that significantly decrease the usability. First of all this threshold seems to be very small as most of changes are not being displayed more. Another point is that PR review feature is affected. You sell a PRO feature to perform a code review via app, however how could it be possible to left a comment if diff isn't displayed? Now I'm able to left comments to a few files in only, while for rest I should use the browser.
I'm very disappointed by this bug :pensive: For me it's about the current approach in general - in each version you introduce some new "features", that later create the usability issues (minor or major) :cry: It's a bit annoying as I feel like a beta tester.. Of course, you quickly fix them, but for couple of weeks we should live with them :( Could you please consider adding an advanced configuration for such things, so we could quickly disable/tune new features if there issues with them? For instance, it would be enough to tune the threshold value to work around the current issue, rather than be unable to perform a code review for a few weeks..
Thank you for your time and work on this project!
|
non_process
|
annoying and blocking too many changes to show fasthub version android version sdk device information manufacturer samsung brand samsung model galaxy tab it seems that you added a threshold in a new version and don t show a diff in pr if file contains many changes this is a horrible decision that significantly decrease the usability first of all this threshold seems to be very small as most of changes are not being displayed more another point is that pr review feature is affected you sell a pro feature to perform a code review via app however how could it be possible to left a comment if diff isn t displayed now i m able to left comments to a few files in only while for rest i should use the browser i m very disappointed by this bug pensive for me it s about the current approach in general in each version you introduce some new features that later create the usability issues minor or major cry it s a bit annoying as i feel like a beta tester of course you quickly fix them but for couple of weeks we should live with them could you please consider adding an advanced configuration for such things so we could quickly disable tune new features if there issues with them for instance it would be enough to tune the threshold value to work around the current issue rather than be unable to perform a code review for a few weeks thank you for your time and work on this project
| 0
|
281,303
| 30,888,605,522
|
IssuesEvent
|
2023-08-04 01:34:02
|
Nivaskumark/kernel_v4.1.15
|
https://api.github.com/repos/Nivaskumark/kernel_v4.1.15
|
reopened
|
WS-2021-0551 (Critical) detected in linuxlinux-4.6
|
Mend: dependency security vulnerability
|
## WS-2021-0551 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.1.15/commit/00db4e8795bcbec692fb60b19160bdd763ad42e3">00db4e8795bcbec692fb60b19160bdd763ad42e3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
nfsd: fix use-after-free due to delegation race
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.8 by commit 148c816f10fd11df27ca6a9b3238cdd42fa72cd3. For more details please see the references link.
<p>Publish Date: 2021-12-19
<p>URL: <a href=https://github.com/gregkh/linux/commit/148c816f10fd11df27ca6a9b3238cdd42fa72cd3>WS-2021-0551</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002600">https://osv.dev/vulnerability/GSD-2021-1002600</a></p>
<p>Release Date: 2021-12-19</p>
<p>Fix Resolution: v5.15.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0551 (Critical) detected in linuxlinux-4.6 - ## WS-2021-0551 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.1.15/commit/00db4e8795bcbec692fb60b19160bdd763ad42e3">00db4e8795bcbec692fb60b19160bdd763ad42e3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
nfsd: fix use-after-free due to delegation race
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.8 by commit 148c816f10fd11df27ca6a9b3238cdd42fa72cd3. For more details please see the references link.
<p>Publish Date: 2021-12-19
<p>URL: <a href=https://github.com/gregkh/linux/commit/148c816f10fd11df27ca6a9b3238cdd42fa72cd3>WS-2021-0551</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002600">https://osv.dev/vulnerability/GSD-2021-1002600</a></p>
<p>Release Date: 2021-12-19</p>
<p>Fix Resolution: v5.15.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws critical detected in linuxlinux ws critical severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details nfsd fix use after free due to delegation race this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
16,810
| 22,060,445,295
|
IssuesEvent
|
2022-05-30 17:02:56
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Terminal close immediately after opening only if I use zsh
|
bug confirmation-pending terminal-process
|
Issue Type: <b>Bug</b>
No typical step required, even with a fresh install and an empty .zshrc file the terminal still close after opening. Even in the developer tools there are no errors. I really don't know what to do...
VS Code version: Code 1.61.2 (Universal) (6cba118ac49a1b88332f312a8f67186f7f3c1643, 2021-10-19T15:49:28.381Z)
OS version: Darwin arm64 20.6.0
Restricted Mode: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Apple M1 (8 x 24)|
|GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled|
|Load (avg)|4, 4, 4|
|Memory (System)|16.00GB (0.62GB free)|
|Process Argv|--crash-reporter-id 658b4308-ce3b-4500-a9ec-e674fc511008|
|Screen Reader|no|
|VM|0%|
</details><details><summary>Extensions (16)</summary>
Extension|Author (truncated)|Version
---|---|---
bracket-pair-colorizer|Coe|1.0.61
doxdocgen|csc|1.3.2
es7-react-js-snippets|dsz|3.1.1
prettier-vscode|esb|9.0.0
auto-rename-tag|for|0.1.9
vscode-jekyll-snippets|gin|0.9.3
solidity|Jua|0.0.134
kite|kit|0.147.0
python|ms-|2021.10.1365161279
vscode-pylance|ms-|2021.10.3
jupyter|ms-|2021.9.1101343141
jupyter-keymap|ms-|1.0.0
jupyter-renderers|ms-|1.0.3
material-icon-theme|PKi|4.10.0
LiveServer|rit|5.6.1
shopify-liquid|sis|2.3.0
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vsreu685:30147344
python383cf:30185419
vspor879:30202332
vspor708:30202333
vspor363:30204092
vswsl492cf:30256860
pythontb:30283811
pythonptprofiler:30281270
vshan820:30294714
vstes263cf:30335440
vscoreces:30384385
pythondataviewer:30285071
pythonvsuse255:30340121
vscod805cf:30301675
pythonvspyt200:30340761
binariesv615:30325510
vsccppwtct:30382698
bridge0708:30335490
pygetstartedt3:30385195
dockerwalkthru:30377721
bridge0723:30353136
pythonrunftest32:30373476
pythonf5test824:30373475
javagetstartedt:30391933
pythonvspyt187:30373474
vsqsis200cf:30386380
vsaa593:30376534
vssld246cf:30386378
```
</details>
<!-- generated by issue reporter -->
|
1.0
|
Terminal close immediately after opening only if I use zsh -
Issue Type: <b>Bug</b>
No typical step required, even with a fresh install and an empty .zshrc file the terminal still close after opening. Even in the developer tools there are no errors. I really don't know what to do...
VS Code version: Code 1.61.2 (Universal) (6cba118ac49a1b88332f312a8f67186f7f3c1643, 2021-10-19T15:49:28.381Z)
OS version: Darwin arm64 20.6.0
Restricted Mode: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Apple M1 (8 x 24)|
|GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled|
|Load (avg)|4, 4, 4|
|Memory (System)|16.00GB (0.62GB free)|
|Process Argv|--crash-reporter-id 658b4308-ce3b-4500-a9ec-e674fc511008|
|Screen Reader|no|
|VM|0%|
</details><details><summary>Extensions (16)</summary>
Extension|Author (truncated)|Version
---|---|---
bracket-pair-colorizer|Coe|1.0.61
doxdocgen|csc|1.3.2
es7-react-js-snippets|dsz|3.1.1
prettier-vscode|esb|9.0.0
auto-rename-tag|for|0.1.9
vscode-jekyll-snippets|gin|0.9.3
solidity|Jua|0.0.134
kite|kit|0.147.0
python|ms-|2021.10.1365161279
vscode-pylance|ms-|2021.10.3
jupyter|ms-|2021.9.1101343141
jupyter-keymap|ms-|1.0.0
jupyter-renderers|ms-|1.0.3
material-icon-theme|PKi|4.10.0
LiveServer|rit|5.6.1
shopify-liquid|sis|2.3.0
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vsreu685:30147344
python383cf:30185419
vspor879:30202332
vspor708:30202333
vspor363:30204092
vswsl492cf:30256860
pythontb:30283811
pythonptprofiler:30281270
vshan820:30294714
vstes263cf:30335440
vscoreces:30384385
pythondataviewer:30285071
pythonvsuse255:30340121
vscod805cf:30301675
pythonvspyt200:30340761
binariesv615:30325510
vsccppwtct:30382698
bridge0708:30335490
pygetstartedt3:30385195
dockerwalkthru:30377721
bridge0723:30353136
pythonrunftest32:30373476
pythonf5test824:30373475
javagetstartedt:30391933
pythonvspyt187:30373474
vsqsis200cf:30386380
vsaa593:30376534
vssld246cf:30386378
```
</details>
<!-- generated by issue reporter -->
|
process
|
terminal close immediately after opening only if i use zsh issue type bug no typical step required even with a fresh install and an empty zshrc file the terminal still close after opening even in the developer tools there are no errors i really don t know what to do vs code version code universal os version darwin restricted mode no system info item value cpus apple x gpu status canvas enabled gpu compositing enabled metal disabled off multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled skia renderer disabled off ok video decode enabled webgl enabled enabled load avg memory system free process argv crash reporter id screen reader no vm extensions extension author truncated version bracket pair colorizer coe doxdocgen csc react js snippets dsz prettier vscode esb auto rename tag for vscode jekyll snippets gin solidity jua kite kit python ms vscode pylance ms jupyter ms jupyter keymap ms jupyter renderers ms material icon theme pki liveserver rit shopify liquid sis a b experiments pythontb pythonptprofiler vscoreces pythondataviewer vsccppwtct dockerwalkthru javagetstartedt
| 1
|
17,378
| 23,200,248,194
|
IssuesEvent
|
2022-08-01 20:38:52
|
Ultimate-Hosts-Blacklist/whitelist
|
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
|
closed
|
[FALSE-POSITIVE?]
|
whitelisting process waiting for Mitch
|
**Domains or links**
Please list any domains and links listed here which you believe are a false positive.
**https://www.airbnb.com.au/**
**More Information**
How did you discover your web site or domain was listed here?
1. Website was hacked
2. Other ..... I see no reason to blacklist a very popular accommodation booking website.
**Have you requested removal from other sources?**
Please include all relevant links to your existing removals / whitelistings.
**Additional context**
Add any other context about the problem here.
:exclamation:
We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
|
1.0
|
[FALSE-POSITIVE?] - **Domains or links**
Please list any domains and links listed here which you believe are a false positive.
**https://www.airbnb.com.au/**
**More Information**
How did you discover your web site or domain was listed here?
1. Website was hacked
2. Other ..... I see no reason to blacklist a very popular accommodation booking website.
**Have you requested removal from other sources?**
Please include all relevant links to your existing removals / whitelistings.
**Additional context**
Add any other context about the problem here.
:exclamation:
We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
|
process
|
domains or links please list any domains and links listed here which you believe are a false positive more information how did you discover your web site or domain was listed here website was hacked other i see no reason to blacklist a very popular accommodation booking website have you requested removal from other sources please include all relevant links to your existing removals whitelistings additional context add any other context about the problem here exclamation we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
| 1
|
85,210
| 24,541,461,734
|
IssuesEvent
|
2022-10-12 04:24:48
|
catboost/catboost
|
https://api.github.com/repos/catboost/catboost
|
closed
|
Error while trying to build
|
build issues
|
Problem:
When I try to run the command
`ya make -r catboost/libs/model_interface`
I got the error:
```
Traceback (most recent call last):
File "devtools/ya/app/init.py", line 657, in configure_exit_interceptor
yield
File "devtools/ya/app/init.py", line 78, in helper
return action(args, **kwargs)
File "devtools/ya/entry/entry.py", line 63, in do_main
res = handler.handle(handler, args, prefix=['ya'])
File "devtools/ya/core/yarg/handler.py", line 222, in handle
return handler.handle(self, args[1:], prefix + [name])
File "devtools/ya/core/yarg/dispatch.py", line 38, in handle
return self.command().handle(root_handler, args, prefix)
File "devtools/ya/core/yarg/handler.py", line 407, in handle
return self._action(params)
File "devtools/ya/app/init.py", line 110, in helper
return action(ctx.params)
File "devtools/ya/build/build_handler.py", line 85, in do_ya_make
builder = ya_make.YaMake(params, app_ctx)
File "devtools/ya/build/ya_make.py", line 924, in init
self.ctx = Context(self.opts, app_ctx=app_ctx, graph=graph, tests=tests, stripped_tests=stripped_tests, configure_errors=configure_errors, make_files=make_files, lite_graph=lite_graph)
File "devtools/ya/build/ya_make.py", line 594, in init
self.graph, self.tests, self.stripped_tests, self.configure_errors, self.make_files = _build_graph_and_tests(self.opts, app_ctx)
File "devtools/ya/build/ya_make.py", line 261, in _build_graph_and_tests
graph, tests, stripped_tests, gh, make_files = lg.build_graph_and_tests(opts, check=True, ev_listener=ev_listener, display=display)
File "devtools/ya/build/graph.py", line 1705, in build_graph_and_tests
return _build_graph_and_tests(opts, check, ev_listener, exit_stack, display)
File "devtools/ya/build/graph.py", line 2076, in _build_graph_and_tests
real_ymake_bin = tools.tool('ymake')
File "devtools/ya/yalibrary/tools/init.py", line 220, in tool
return toolchain.find(name, with_params, for_platform, cache=cache)
File "devtools/ya/yalibrary/tools/init.py", line 155, in find
cur_bottle = _bottle(location['bottle'], for_platform)
File "devtools/ya/yalibrary/tools/init.py", line 93, in _bottle
return _bottler().get(name, for_platform)
File "devtools/ya/yalibrary/tools/init.py", line 82, in get
return _Bottle(value['formula'], value.get('executable'), for_platform)
File "devtools/ya/yalibrary/tools/init.py", line 40, in init
self.fetcher = yalibrary.fetcher.ToolChainFetcher(core.config.tool_root(toolscache_version()), self.formula, for_platform)
File "devtools/ya/yalibrary/fetcher/init.py", line 586, in ToolChainFetcher
impl = _ToolChainSandboxFetcher(root, formula["sandbox_id"], for_platform)
TypeError: string indices must be integers, not str
```
catboost version: latest (07.10.22)
Operating System: Windows 10 x64
CPU: Intel Core i3-1125G4
|
1.0
|
Error while trying to build - Problem:
When I try to run the command
`ya make -r catboost/libs/model_interface`
I got the error:
```
Traceback (most recent call last):
File "devtools/ya/app/init.py", line 657, in configure_exit_interceptor
yield
File "devtools/ya/app/init.py", line 78, in helper
return action(args, **kwargs)
File "devtools/ya/entry/entry.py", line 63, in do_main
res = handler.handle(handler, args, prefix=['ya'])
File "devtools/ya/core/yarg/handler.py", line 222, in handle
return handler.handle(self, args[1:], prefix + [name])
File "devtools/ya/core/yarg/dispatch.py", line 38, in handle
return self.command().handle(root_handler, args, prefix)
File "devtools/ya/core/yarg/handler.py", line 407, in handle
return self._action(params)
File "devtools/ya/app/init.py", line 110, in helper
return action(ctx.params)
File "devtools/ya/build/build_handler.py", line 85, in do_ya_make
builder = ya_make.YaMake(params, app_ctx)
File "devtools/ya/build/ya_make.py", line 924, in init
self.ctx = Context(self.opts, app_ctx=app_ctx, graph=graph, tests=tests, stripped_tests=stripped_tests, configure_errors=configure_errors, make_files=make_files, lite_graph=lite_graph)
File "devtools/ya/build/ya_make.py", line 594, in init
self.graph, self.tests, self.stripped_tests, self.configure_errors, self.make_files = _build_graph_and_tests(self.opts, app_ctx)
File "devtools/ya/build/ya_make.py", line 261, in _build_graph_and_tests
graph, tests, stripped_tests, gh, make_files = lg.build_graph_and_tests(opts, check=True, ev_listener=ev_listener, display=display)
File "devtools/ya/build/graph.py", line 1705, in build_graph_and_tests
return _build_graph_and_tests(opts, check, ev_listener, exit_stack, display)
File "devtools/ya/build/graph.py", line 2076, in _build_graph_and_tests
real_ymake_bin = tools.tool('ymake')
File "devtools/ya/yalibrary/tools/init.py", line 220, in tool
return toolchain.find(name, with_params, for_platform, cache=cache)
File "devtools/ya/yalibrary/tools/init.py", line 155, in find
cur_bottle = _bottle(location['bottle'], for_platform)
File "devtools/ya/yalibrary/tools/init.py", line 93, in _bottle
return _bottler().get(name, for_platform)
File "devtools/ya/yalibrary/tools/init.py", line 82, in get
return _Bottle(value['formula'], value.get('executable'), for_platform)
File "devtools/ya/yalibrary/tools/init.py", line 40, in init
self.fetcher = yalibrary.fetcher.ToolChainFetcher(core.config.tool_root(toolscache_version()), self.formula, for_platform)
File "devtools/ya/yalibrary/fetcher/init.py", line 586, in ToolChainFetcher
impl = _ToolChainSandboxFetcher(root, formula["sandbox_id"], for_platform)
TypeError: string indices must be integers, not str
```
catboost version: latest (07.10.22)
Operating System: Windows 10 x64
CPU: Intel Core i3-1125G4
|
non_process
|
error while trying to build problem when i try to run the command ya make r catboost libs model interface i got the error traceback most recent call last file devtools ya app init py line in configure exit interceptor yield file devtools ya app init py line in helper return action args kwargs file devtools ya entry entry py line in do main res handler handle handler args prefix file devtools ya core yarg handler py line in handle return handler handle self args prefix file devtools ya core yarg dispatch py line in handle return self command handle root handler args prefix file devtools ya core yarg handler py line in handle return self action params file devtools ya app init py line in helper return action ctx params file devtools ya build build handler py line in do ya make builder ya make yamake params app ctx file devtools ya build ya make py line in init self ctx context self opts app ctx app ctx graph graph tests tests stripped tests stripped tests configure errors configure errors make files make files lite graph lite graph file devtools ya build ya make py line in init self graph self tests self stripped tests self configure errors self make files build graph and tests self opts app ctx file devtools ya build ya make py line in build graph and tests graph tests stripped tests gh make files lg build graph and tests opts check true ev listener ev listener display display file devtools ya build graph py line in build graph and tests return build graph and tests opts check ev listener exit stack display file devtools ya build graph py line in build graph and tests real ymake bin tools tool ymake file devtools ya yalibrary tools init py line in tool return toolchain find name with params for platform cache cache file devtools ya yalibrary tools init py line in find cur bottle bottle location for platform file devtools ya yalibrary tools init py line in bottle return bottler get name for platform file devtools ya yalibrary tools init py line in get return bottle value value get executable for platform file devtools ya yalibrary tools init py line in init self fetcher yalibrary fetcher toolchainfetcher core config tool root toolscache version self formula for platform file devtools ya yalibrary fetcher init py line in toolchainfetcher impl toolchainsandboxfetcher root formula for platform typeerror string indices must be integers not str catboost version latest operating system windows cpu intel core
| 0
|
453,598
| 13,085,208,943
|
IssuesEvent
|
2020-08-02 00:48:13
|
SpencerTSterling/RatHutProductManager
|
https://api.github.com/repos/SpencerTSterling/RatHutProductManager
|
closed
|
Create a Database class
|
enhancement high priority
|
- Add a class to the project ex. ProductDb
- Code method stubds ex. Add(), Update(), Delete()
``` csharp
throw new NotImplementedException();
```
|
1.0
|
Create a Database class - - Add a class to the project ex. ProductDb
- Code method stubds ex. Add(), Update(), Delete()
``` csharp
throw new NotImplementedException();
```
|
non_process
|
create a database class add a class to the project ex productdb code method stubds ex add update delete csharp throw new notimplementedexception
| 0
|
3,180
| 6,256,013,537
|
IssuesEvent
|
2017-07-14 08:59:03
|
rogerthat-platform/rogerthat-android-client
|
https://api.github.com/repos/rogerthat-platform/rogerthat-android-client
|
closed
|
IOException in startScanningForQRCodes
|
priority_major process_wontfix type_bug
|
```
2017-07-03 07:56:22 UTC
(16 min ago)
user008a407f635446b5a4e78d282371d3da@***:osa-loyalty
Count: 38 - Version: 0.2732
```
```
description: Bug!
Unexpected IOException initializing TextureView camera
errorMessage: java.io.IOException: setPreviewTexture failed
at android.hardware.Camera.setPreviewTexture(Native Method)
at com.google.zxing.client.android.camera.QRCodeScannerTextureViewCameraManager.openDriver(QRCodeScannerTextureViewCameraManager.java:60)
at com.mobicage.rogerthat.plugins.friends.QRCodeScannerTextureView.startScanningForQRCodes(QRCodeScannerTextureView.java:80)
at com.mobicage.rogerthat.plugins.friends.QRCodeScanner.onResume(QRCodeScanner.java:132)
at com.mobicage.rogerthat.plugins.friends.ActionScreenActivity.onResume(ActionScreenActivity.java:900)
at android.app.Instrumentation.callActivityOnResume(Instrumentation.java:1285)
at android.app.Activity.performResume(Activity.java:6730)
at android.app.ActivityThread.performResumeActivity(ActivityThread.java:4025)
at android.app.ActivityThread.handleResumeActivity(ActivityThread.java:4105)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1779)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:145)
at android.app.ActivityThread.main(ActivityThread.java:6939)
at java.lang.reflect.Method.invoke(Native Method)
at java.lang.reflect.Method.invoke(Method.java:372)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1404)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1199)
mobicageVersion: 0.2732
occurenceCount: 38
platform: 1
platformVersion: samsung/gtexswifixx/gtexswifi:5.1.1/LMY47V/T280XXU0AQA4:user/release-keys (-) 22 (-) SM-T280
```
|
1.0
|
IOException in startScanningForQRCodes - ```
2017-07-03 07:56:22 UTC
(16 min ago)
user008a407f635446b5a4e78d282371d3da@***:osa-loyalty
Count: 38 - Version: 0.2732
```
```
description: Bug!
Unexpected IOException initializing TextureView camera
errorMessage: java.io.IOException: setPreviewTexture failed
at android.hardware.Camera.setPreviewTexture(Native Method)
at com.google.zxing.client.android.camera.QRCodeScannerTextureViewCameraManager.openDriver(QRCodeScannerTextureViewCameraManager.java:60)
at com.mobicage.rogerthat.plugins.friends.QRCodeScannerTextureView.startScanningForQRCodes(QRCodeScannerTextureView.java:80)
at com.mobicage.rogerthat.plugins.friends.QRCodeScanner.onResume(QRCodeScanner.java:132)
at com.mobicage.rogerthat.plugins.friends.ActionScreenActivity.onResume(ActionScreenActivity.java:900)
at android.app.Instrumentation.callActivityOnResume(Instrumentation.java:1285)
at android.app.Activity.performResume(Activity.java:6730)
at android.app.ActivityThread.performResumeActivity(ActivityThread.java:4025)
at android.app.ActivityThread.handleResumeActivity(ActivityThread.java:4105)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1779)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:145)
at android.app.ActivityThread.main(ActivityThread.java:6939)
at java.lang.reflect.Method.invoke(Native Method)
at java.lang.reflect.Method.invoke(Method.java:372)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1404)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1199)
mobicageVersion: 0.2732
occurenceCount: 38
platform: 1
platformVersion: samsung/gtexswifixx/gtexswifi:5.1.1/LMY47V/T280XXU0AQA4:user/release-keys (-) 22 (-) SM-T280
```
|
process
|
ioexception in startscanningforqrcodes utc min ago osa loyalty count version description bug unexpected ioexception initializing textureview camera errormessage java io ioexception setpreviewtexture failed at android hardware camera setpreviewtexture native method at com google zxing client android camera qrcodescannertextureviewcameramanager opendriver qrcodescannertextureviewcameramanager java at com mobicage rogerthat plugins friends qrcodescannertextureview startscanningforqrcodes qrcodescannertextureview java at com mobicage rogerthat plugins friends qrcodescanner onresume qrcodescanner java at com mobicage rogerthat plugins friends actionscreenactivity onresume actionscreenactivity java at android app instrumentation callactivityonresume instrumentation java at android app activity performresume activity java at android app activitythread performresumeactivity activitythread java at android app activitythread handleresumeactivity activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java mobicageversion occurencecount platform platformversion samsung gtexswifixx gtexswifi user release keys sm
| 1
|
4,362
| 7,260,514,351
|
IssuesEvent
|
2018-02-18 10:53:27
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE][processing] New algorithm to extract specific nodes
|
Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/82312e10e472fb91a5b5c0253456130dea86521e by nyalldawson
This algorithm allows you to extract specific nodes from geometries.
Eg you can extract the first or last node in the geometry.
The algorithm accepts a comma separated list of node indices to
extract, eg 0 = first node, 1 = second node, etc. Negative indices
can be used to extract nodes from the end of the geometry. Eg
-1 = last node, -2 = second last node.
|
1.0
|
[FEATURE][processing] New algorithm to extract specific nodes - Original commit: https://github.com/qgis/QGIS/commit/82312e10e472fb91a5b5c0253456130dea86521e by nyalldawson
This algorithm allows you to extract specific nodes from geometries.
Eg you can extract the first or last node in the geometry.
The algorithm accepts a comma separated list of node indices to
extract, eg 0 = first node, 1 = second node, etc. Negative indices
can be used to extract nodes from the end of the geometry. Eg
-1 = last node, -2 = second last node.
|
process
|
new algorithm to extract specific nodes original commit by nyalldawson this algorithm allows you to extract specific nodes from geometries eg you can extract the first or last node in the geometry the algorithm accepts a comma separated list of node indices to extract eg first node second node etc negative indices can be used to extract nodes from the end of the geometry eg last node second last node
| 1
|
7,571
| 10,684,626,204
|
IssuesEvent
|
2019-10-22 10:54:13
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
Proposed obsoletion multi-organism transport
|
multi-species process obsoletion
|
Dear all,
The proposal has been made to obsolete the following terms:
GO:1902537 multi-organism pinocytosis
GO:1990967 multi-organism toxin transport
GO:1902586 multi-organism intercellular transport
GO:1902588 multi-organism plasmodesmata-mediated intercellular transport
The reason for the obsoletion is that only host proteins take part in these processes, so it is NOT a multi-organism process.
There are no annotations, no mappings to those terms. Those terms are not present in any slims.
Any comments can be added to the issue: https://github.com/geneontology/go-ontology/issues
We are opening a comment period for this proposed obsoletion. We’d like to proceed and obsolete this term on October 28th, 2019. Unless objections are received by October 28th 2019, we will assume that you agree to this change.
Thanks, Pascale
|
1.0
|
Proposed obsoletion multi-organism transport - Dear all,
The proposal has been made to obsolete the following terms:
GO:1902537 multi-organism pinocytosis
GO:1990967 multi-organism toxin transport
GO:1902586 multi-organism intercellular transport
GO:1902588 multi-organism plasmodesmata-mediated intercellular transport
The reason for the obsoletion is that only host proteins take part in these processes, so it is NOT a multi-organism process.
There are no annotations, no mappings to those terms. Those terms are not present in any slims.
Any comments can be added to the issue: https://github.com/geneontology/go-ontology/issues
We are opening a comment period for this proposed obsoletion. We’d like to proceed and obsolete this term on October 28th, 2019. Unless objections are received by October 28th 2019, we will assume that you agree to this change.
Thanks, Pascale
|
process
|
proposed obsoletion multi organism transport dear all the proposal has been made to obsolete the following terms go multi organism pinocytosis go multi organism toxin transport go multi organism intercellular transport go multi organism plasmodesmata mediated intercellular transport the reason for the obsoletion is that only host proteins take part in these processes so it is not a multi organism process there are no annotations no mappings to those terms those terms are not present in any slims any comments can be added to the issue we are opening a comment period for this proposed obsoletion we’d like to proceed and obsolete this term on october unless objections are received by october we will assume that you agree to this change thanks pascale
| 1
|
70,019
| 30,531,613,739
|
IssuesEvent
|
2023-07-19 14:36:28
|
Ocelot-Social-Community/Ocelot-Social
|
https://api.github.com/repos/Ocelot-Social-Community/Ocelot-Social
|
closed
|
🚀 [Feature][Webapp] Chat room overview: last message text should be a line with ~50chars
|
feature service: webapp features_chat chat
|
<!-- You can find the latest issue templates here https://github.com/ulfgebhardt/issue-templates -->
## 🚀 Feature
<!-- Give a short summary of the Feature. Use Screenshots if you want. -->

## 🤖 ToDo
- [ ] Compute the last message teaser
- [ ] Update chat room last message slot
|
1.0
|
🚀 [Feature][Webapp] Chat room overview: last message text should be a line with ~50chars - <!-- You can find the latest issue templates here https://github.com/ulfgebhardt/issue-templates -->
## 🚀 Feature
<!-- Give a short summary of the Feature. Use Screenshots if you want. -->

## 🤖 ToDo
- [ ] Compute the last message teaser
- [ ] Update chat room last message slot
|
non_process
|
🚀 chat room overview last message text should be a line with 🚀 feature 🤖 todo compute the last message teaser update chat room last message slot
| 0
|
15,390
| 19,571,819,885
|
IssuesEvent
|
2022-01-04 10:51:13
|
Kernem/FeRSS-Core
|
https://api.github.com/repos/Kernem/FeRSS-Core
|
closed
|
Filter RSS content by length
|
post-processing filter
|
RSS content should be filterable by length, ideally fetching and processing should stop processing some content if it falls outside of the length
|
1.0
|
Filter RSS content by length - RSS content should be filterable by length, ideally fetching and processing should stop processing some content if it falls outside of the length
|
process
|
filter rss content by length rss content should be filterable by length ideally fetching and processing should stop processing some content if it falls outside of the length
| 1
|
495,623
| 14,285,286,249
|
IssuesEvent
|
2020-11-23 13:43:04
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.dailymail.co.uk - video or audio doesn't play
|
browser-firefox engine-gecko ml-needsdiagnosis-false priority-important
|
<!-- @browser: Firefox 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/62309 -->
**URL**: https://www.dailymail.co.uk/ushome/index.html
**Browser / Version**: Firefox 84.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video appears, but will not permit it to play. Also, for some news stories, at times there is no accompanying photograph on the main page.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.dailymail.co.uk - video or audio doesn't play - <!-- @browser: Firefox 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/62309 -->
**URL**: https://www.dailymail.co.uk/ushome/index.html
**Browser / Version**: Firefox 84.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video appears, but will not permit it to play. Also, for some news stories, at times there is no accompanying photograph on the main page.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
video or audio doesn t play url browser version firefox operating system windows tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce video appears but will not permit it to play also for some news stories at times there is no accompanying photograph on the main page browser configuration none from with ❤️
| 0
|
4,563
| 7,393,696,829
|
IssuesEvent
|
2018-03-17 00:47:07
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Possible to import XML doc of Azure agent IPs?
|
app-service-web cxp in-process product-question triaged
|
We need to white-list the Azure IPs used for running performance tests against our web app. We cannot find a definitive answer on which IPs these are. So, my though was to import the Azure IP data ranges XML. Thoughts?
https://www.microsoft.com/en-us/download/confirmation.aspx?id=41653
Joey
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d90bbf62-44f9-30a6-d6ea-bbf776f6efc5
* Version Independent ID: b3961eac-6c70-b345-f48c-e4c0ec67059c
* Content: [Azure App Service IP Restrictions](https://docs.microsoft.com/en-us/azure/app-service/app-service-ip-restrictions)
* Content Source: [articles/app-service/app-service-ip-restrictions.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-ip-restrictions.md)
* Service: **app-service-web**
* GitHub Login: @btardif
* Microsoft Alias: **byvinyal**
|
1.0
|
Possible to import XML doc of Azure agent IPs? - We need to white-list the Azure IPs used for running performance tests against our web app. We cannot find a definitive answer on which IPs these are. So, my though was to import the Azure IP data ranges XML. Thoughts?
https://www.microsoft.com/en-us/download/confirmation.aspx?id=41653
Joey
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d90bbf62-44f9-30a6-d6ea-bbf776f6efc5
* Version Independent ID: b3961eac-6c70-b345-f48c-e4c0ec67059c
* Content: [Azure App Service IP Restrictions](https://docs.microsoft.com/en-us/azure/app-service/app-service-ip-restrictions)
* Content Source: [articles/app-service/app-service-ip-restrictions.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-ip-restrictions.md)
* Service: **app-service-web**
* GitHub Login: @btardif
* Microsoft Alias: **byvinyal**
|
process
|
possible to import xml doc of azure agent ips we need to white list the azure ips used for running performance tests against our web app we cannot find a definitive answer on which ips these are so my though was to import the azure ip data ranges xml thoughts joey document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service web github login btardif microsoft alias byvinyal
| 1
|
9,698
| 12,700,216,845
|
IssuesEvent
|
2020-06-22 15:59:00
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
System.Diagnostic.Process.Start(...) does not Support mailto: protocoll
|
area-System.Diagnostics.Process untriaged
|
Porting a .Net Framework App to .Net Core 3.0, I was not able to invoke the default mail client with the `mailto:` protocoll using `System.Diagnostics.Process.Start(...)`.
.Net Core `3.0.100-preview5-011568`
Windows 10 1809
Outlook 2016
Sampe code to reproduce
```c#
var body = "This is a body of a message";
var recipients = String.Join(",", "to@user.com", "alsoTo@user.com");
string mailto = $"mailto:{recipients}?Subject={"Subject of message"}&Body={body}";
mailto = Uri.EscapeUriString(mailto);
System.Diagnostics.Process.Start(mailto);
```
The Exception that occures: `System.ComponentModel.Win32Exception: 'Das System kann die angegebene Datei nicht finden.'` (Translated `System can not find the file` not sure if that is the actuall english error message)
|
1.0
|
System.Diagnostic.Process.Start(...) does not Support mailto: protocoll -
Porting a .Net Framework App to .Net Core 3.0, I was not able to invoke the default mail client with the `mailto:` protocoll using `System.Diagnostics.Process.Start(...)`.
.Net Core `3.0.100-preview5-011568`
Windows 10 1809
Outlook 2016
Sampe code to reproduce
```c#
var body = "This is a body of a message";
var recipients = String.Join(",", "to@user.com", "alsoTo@user.com");
string mailto = $"mailto:{recipients}?Subject={"Subject of message"}&Body={body}";
mailto = Uri.EscapeUriString(mailto);
System.Diagnostics.Process.Start(mailto);
```
The Exception that occures: `System.ComponentModel.Win32Exception: 'Das System kann die angegebene Datei nicht finden.'` (Translated `System can not find the file` not sure if that is the actuall english error message)
|
process
|
system diagnostic process start does not support mailto protocoll porting a net framework app to net core i was not able to invoke the default mail client with the mailto protocoll using system diagnostics process start net core windows outlook sampe code to reproduce c var body this is a body of a message var recipients string join to user com alsoto user com string mailto mailto recipients subject subject of message body body mailto uri escapeuristring mailto system diagnostics process start mailto the exception that occures system componentmodel das system kann die angegebene datei nicht finden translated system can not find the file not sure if that is the actuall english error message
| 1
|
12,813
| 15,082,920,292
|
IssuesEvent
|
2021-02-05 15:10:59
|
cseelhoff/RimThreaded
|
https://api.github.com/repos/cseelhoff/RimThreaded
|
opened
|
"Insulation (Continued)" frequent errors from doors
|
1.3.0 - 1.4.0 Bug Confirmed Bug Mod Incompatibility Reproducible
|
**Describe the bug**
IMPORTANT: Please first search existing bugs to ensure you are not creating a duplicate bug report.
errors with insulation mod enabled
**To Reproduce (VERY IMPORTANT)**
Steps to reproduce the behavior:
1. Load a save
2. See errors
**Error Log**
https://gist.github.com/1163af431c358ee45354c634418805a2
**Mod List**
see log
https://steamcommunity.com/sharedfiles/filedetails/?id=2196454890&searchtext=insulation
**Screenshots**
* NA
|
True
|
"Insulation (Continued)" frequent errors from doors - **Describe the bug**
IMPORTANT: Please first search existing bugs to ensure you are not creating a duplicate bug report.
errors with insulation mod enabled
**To Reproduce (VERY IMPORTANT)**
Steps to reproduce the behavior:
1. Load a save
2. See errors
**Error Log**
https://gist.github.com/1163af431c358ee45354c634418805a2
**Mod List**
see log
https://steamcommunity.com/sharedfiles/filedetails/?id=2196454890&searchtext=insulation
**Screenshots**
* NA
|
non_process
|
insulation continued frequent errors from doors describe the bug important please first search existing bugs to ensure you are not creating a duplicate bug report errors with insulation mod enabled to reproduce very important steps to reproduce the behavior load a save see errors error log mod list see log screenshots na
| 0
|
22,448
| 31,168,793,016
|
IssuesEvent
|
2023-08-16 22:18:58
|
NCAR/ucomp-pipeline
|
https://api.github.com/repos/NCAR/ucomp-pipeline
|
opened
|
Reprocess test dates before 0.5.x release/reprocessing
|
needs testing process
|
Reprocess the following dates:
- [ ] 20210719
- [ ] 20220325
- [ ] 20220523
- [ ] 20220712
Check the following:
- no level 2 artifacts
- good threshold masking in level 2 products
- good field radius masking in level 2 PNGs
|
1.0
|
Reprocess test dates before 0.5.x release/reprocessing - Reprocess the following dates:
- [ ] 20210719
- [ ] 20220325
- [ ] 20220523
- [ ] 20220712
Check the following:
- no level 2 artifacts
- good threshold masking in level 2 products
- good field radius masking in level 2 PNGs
|
process
|
reprocess test dates before x release reprocessing reprocess the following dates check the following no level artifacts good threshold masking in level products good field radius masking in level pngs
| 1
|
191,679
| 14,595,180,811
|
IssuesEvent
|
2020-12-20 10:10:17
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
codeamp/circuit: plugins/codeamp/graphql/helpers_test.go; 3 LoC
|
fresh test tiny
|
Found a possible issue in [codeamp/circuit](https://www.github.com/codeamp/circuit) at [plugins/codeamp/graphql/helpers_test.go](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to projectEnvironment at line 637 may start a goroutine
[Click here to see the code in its original context.](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, projectEnvironment := range projectEnvironments {
helper.Resolver.DB.Unscoped().Delete(&projectEnvironment)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 06371150b7aaaf104d41907c6f432a795238bf3c
|
1.0
|
codeamp/circuit: plugins/codeamp/graphql/helpers_test.go; 3 LoC -
Found a possible issue in [codeamp/circuit](https://www.github.com/codeamp/circuit) at [plugins/codeamp/graphql/helpers_test.go](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to projectEnvironment at line 637 may start a goroutine
[Click here to see the code in its original context.](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, projectEnvironment := range projectEnvironments {
helper.Resolver.DB.Unscoped().Delete(&projectEnvironment)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 06371150b7aaaf104d41907c6f432a795238bf3c
|
non_process
|
codeamp circuit plugins codeamp graphql helpers test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to projectenvironment at line may start a goroutine click here to show the line s of go which triggered the analyzer go for projectenvironment range projectenvironments helper resolver db unscoped delete projectenvironment leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
36,102
| 9,745,265,560
|
IssuesEvent
|
2019-06-03 09:12:54
|
groupe-sii/ogham
|
https://api.github.com/repos/groupe-sii/ogham
|
opened
|
Provide configurers for mainly used service providers
|
API Builder P4 enhancement
|
Provide a simple automatic way to configure email sending through Gmail, Yahoo, ...
It could be something like this with fluent API:
```java
.email()
.gmail()
.account()
.username()
.applicationPassword()
```
An using properties:
```
ogham.email.gmail.account.username=
ogham.email.gmail.account.application-password=
```
And nothing else to do. It will automatically configure correctly the SMTP configuration
|
1.0
|
Provide configurers for mainly used service providers - Provide a simple automatic way to configure email sending through Gmail, Yahoo, ...
It could be something like this with fluent API:
```java
.email()
.gmail()
.account()
.username()
.applicationPassword()
```
An using properties:
```
ogham.email.gmail.account.username=
ogham.email.gmail.account.application-password=
```
And nothing else to do. It will automatically configure correctly the SMTP configuration
|
non_process
|
provide configurers for mainly used service providers provide a simple automatic way to configure email sending through gmail yahoo it could be something like this with fluent api java email gmail account username applicationpassword an using properties ogham email gmail account username ogham email gmail account application password and nothing else to do it will automatically configure correctly the smtp configuration
| 0
|
18,205
| 24,260,500,334
|
IssuesEvent
|
2022-09-27 22:06:51
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
concurrent.futures.as_completed() installs waiters for already completed Futures
|
performance stdlib pending expert-multiprocessing
|
BPO | [20297](https://bugs.python.org/issue20297)
--- | :---
Nosy | @brianquinlan, @mdickinson, @vstinner
Files | <li>[as_completed_proposed.py](https://bugs.python.org/file33590/as_completed_proposed.py "Uploaded as text/plain at 2014-01-21.15:19:28 by glangford")</li><li>[as_completed_proposed.patch](https://bugs.python.org/file33654/as_completed_proposed.patch "Uploaded as text/plain at 2014-01-23.13:48:02 by @vstinner")</li><li>[test_dupfuture.py](https://bugs.python.org/file33657/test_dupfuture.py "Uploaded as text/plain at 2014-01-23.14:43:21 by glangford")</li>
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2014-01-18.21:19:35.870>
labels = ['library', 'performance']
title = 'concurrent.futures.as_completed() installs waiters for already completed Futures'
updated_at = <Date 2014-07-18.16:46:17.165>
user = 'https://bugs.python.org/glangford'
```
bugs.python.org fields:
```python
activity = <Date 2014-07-18.16:46:17.165>
actor = 'glangford'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = ['Library (Lib)']
creation = <Date 2014-01-18.21:19:35.870>
creator = 'glangford'
dependencies = []
files = ['33590', '33654', '33657']
hgrepos = []
issue_num = 20297
keywords = ['patch']
message_count = 7.0
messages = ['208418', '208593', '208652', '208921', '208942', '208943', '208974']
nosy_count = 3.0
nosy_names = ['bquinlan', 'mark.dickinson', 'vstinner']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'performance'
url = 'https://bugs.python.org/issue20297'
versions = ['Python 3.3', 'Python 3.4']
```
</p></details>
|
1.0
|
concurrent.futures.as_completed() installs waiters for already completed Futures - BPO | [20297](https://bugs.python.org/issue20297)
--- | :---
Nosy | @brianquinlan, @mdickinson, @vstinner
Files | <li>[as_completed_proposed.py](https://bugs.python.org/file33590/as_completed_proposed.py "Uploaded as text/plain at 2014-01-21.15:19:28 by glangford")</li><li>[as_completed_proposed.patch](https://bugs.python.org/file33654/as_completed_proposed.patch "Uploaded as text/plain at 2014-01-23.13:48:02 by @vstinner")</li><li>[test_dupfuture.py](https://bugs.python.org/file33657/test_dupfuture.py "Uploaded as text/plain at 2014-01-23.14:43:21 by glangford")</li>
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2014-01-18.21:19:35.870>
labels = ['library', 'performance']
title = 'concurrent.futures.as_completed() installs waiters for already completed Futures'
updated_at = <Date 2014-07-18.16:46:17.165>
user = 'https://bugs.python.org/glangford'
```
bugs.python.org fields:
```python
activity = <Date 2014-07-18.16:46:17.165>
actor = 'glangford'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = ['Library (Lib)']
creation = <Date 2014-01-18.21:19:35.870>
creator = 'glangford'
dependencies = []
files = ['33590', '33654', '33657']
hgrepos = []
issue_num = 20297
keywords = ['patch']
message_count = 7.0
messages = ['208418', '208593', '208652', '208921', '208942', '208943', '208974']
nosy_count = 3.0
nosy_names = ['bquinlan', 'mark.dickinson', 'vstinner']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'performance'
url = 'https://bugs.python.org/issue20297'
versions = ['Python 3.3', 'Python 3.4']
```
</p></details>
|
process
|
concurrent futures as completed installs waiters for already completed futures bpo nosy brianquinlan mdickinson vstinner files uploaded as text plain at by glangford uploaded as text plain at by vstinner uploaded as text plain at by glangford note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title concurrent futures as completed installs waiters for already completed futures updated at user bugs python org fields python activity actor glangford assignee none closed false closed date none closer none components creation creator glangford dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type performance url versions
| 1
|
432,873
| 30,297,256,998
|
IssuesEvent
|
2023-07-10 00:41:59
|
RE-M4/PV-Final-2023
|
https://api.github.com/repos/RE-M4/PV-Final-2023
|
closed
|
Agregar documentación al controlador Testimonio
|
documentation
|
-Se debe agregar documentación a los métodos del controlador y eliminar líneas de código innecesarias de haberlas.
|
1.0
|
Agregar documentación al controlador Testimonio - -Se debe agregar documentación a los métodos del controlador y eliminar líneas de código innecesarias de haberlas.
|
non_process
|
agregar documentación al controlador testimonio se debe agregar documentación a los métodos del controlador y eliminar líneas de código innecesarias de haberlas
| 0
|
12,984
| 15,356,566,942
|
IssuesEvent
|
2021-03-01 12:36:17
|
GoogleCloudPlatform/dotnet-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
|
closed
|
Asset: Quickstart tests sometimes timing out
|
api: cloudasset priority: p1 samples type: process
|
Sample [CI output](https://source.cloud.google.com/results/invocations/8273bb2e-f90b-4617-92b3-6aaf17a0279c/targets/github%2Fdotnet-docs-samples%2Fasset%2Fquickstart%2FExportAssetsTest/tests)
This might just be a matter of increasing timeouts.
|
1.0
|
Asset: Quickstart tests sometimes timing out - Sample [CI output](https://source.cloud.google.com/results/invocations/8273bb2e-f90b-4617-92b3-6aaf17a0279c/targets/github%2Fdotnet-docs-samples%2Fasset%2Fquickstart%2FExportAssetsTest/tests)
This might just be a matter of increasing timeouts.
|
process
|
asset quickstart tests sometimes timing out sample this might just be a matter of increasing timeouts
| 1
|
10,678
| 13,462,471,860
|
IssuesEvent
|
2020-09-09 16:07:06
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Add to generation scripts for BCO
|
Format - RDF Process - dismissed priority
|
BCO always imports from Darwin Core. Create the dwc.owl file from the normative rdf using the scripts.
|
1.0
|
Add to generation scripts for BCO - BCO always imports from Darwin Core. Create the dwc.owl file from the normative rdf using the scripts.
|
process
|
add to generation scripts for bco bco always imports from darwin core create the dwc owl file from the normative rdf using the scripts
| 1
|
14,900
| 2,610,838,157
|
IssuesEvent
|
2015-02-26 22:46:17
|
tsgrp/ActiveWizard
|
https://api.github.com/repos/tsgrp/ActiveWizard
|
closed
|
Workflow rules are not evaluating properly
|
High Priority HTML5 issue
|
See the screencam below. It looks like the workflow summary page is not displaying the correct users for the Initial QA role.

I see two issues:
- [ ] The workflow summary page is displaying the wrong user for the IQA role
- [ ] The reason for the approver being selected is wrong (see below)
In the screenshot below, it looks like the high level rule is being used to determine the approver. The IQA role should trip with:
- High level rule - Priority of Change is not Low
- User level rule - Priority of Change is Medium
- User is Max

@mikeblum @dgrumieaux - Not sure if this is a problem for our current client's form, but we'll definitely need this for the HPI 2.3 release since we primarily demo with the Simple CR form.
|
1.0
|
Workflow rules are not evaluating properly - See the screencam below. It looks like the workflow summary page is not displaying the correct users for the Initial QA role.

I see two issues:
- [ ] The workflow summary page is displaying the wrong user for the IQA role
- [ ] The reason for the approver being selected is wrong (see below)
In the screenshot below, it looks like the high level rule is being used to determine the approver. The IQA role should trip with:
- High level rule - Priority of Change is not Low
- User level rule - Priority of Change is Medium
- User is Max

@mikeblum @dgrumieaux - Not sure if this is a problem for our current client's form, but we'll definitely need this for the HPI 2.3 release since we primarily demo with the Simple CR form.
|
non_process
|
workflow rules are not evaluating properly see the screencam below it looks like the workflow summary page is not displaying the correct users for the initial qa role i see two issues the workflow summary page is displaying the wrong user for the iqa role the reason for the approver being selected is wrong see below in the screenshot below it looks like the high level rule is being used to determine the approver the iqa role should trip with high level rule priority of change is not low user level rule priority of change is medium user is max mikeblum dgrumieaux not sure if this is a problem for our current client s form but we ll definitely need this for the hpi release since we primarily demo with the simple cr form
| 0
|
741,042
| 25,777,767,815
|
IssuesEvent
|
2022-12-09 13:27:06
|
bounswe/bounswe2022group4
|
https://api.github.com/repos/bounswe/bounswe2022group4
|
closed
|
Frontend: Update Create Comment UI
|
Category - To Do Priority - High Status: In Progress whom: individual Difficulty - Hard Language - React.js Team - Frontend
|
I have implemented a create comment structure that only allow user to provide body. Since backend has changed now i need to implement a structure that allow user to provide upvote a comment, downvote a comment and delete a comment.
Steps:
1) Make research on material ui modal structure
2) Implementing a new UI for the Create Comment Component
3) Add new fields with functionalities in Create Comment Component
Reviewer: @BeratDamar
Deadline: 04.12.2022 23.59
|
1.0
|
Frontend: Update Create Comment UI - I have implemented a create comment structure that only allow user to provide body. Since backend has changed now i need to implement a structure that allow user to provide upvote a comment, downvote a comment and delete a comment.
Steps:
1) Make research on material ui modal structure
2) Implementing a new UI for the Create Comment Component
3) Add new fields with functionalities in Create Comment Component
Reviewer: @BeratDamar
Deadline: 04.12.2022 23.59
|
non_process
|
frontend update create comment ui i have implemented a create comment structure that only allow user to provide body since backend has changed now i need to implement a structure that allow user to provide upvote a comment downvote a comment and delete a comment steps make research on material ui modal structure implementing a new ui for the create comment component add new fields with functionalities in create comment component reviewer beratdamar deadline
| 0
|
1,678
| 4,314,001,904
|
IssuesEvent
|
2016-07-22 12:59:34
|
pelias/api
|
https://api.github.com/repos/pelias/api
|
closed
|
Move back to upstream cluster2 package if our PR is merged
|
processed
|
Update package.json to remove the `git://` url used as an interim solution until the upstream PR is closed, see commit: https://github.com/pelias/api/commit/e3efeb66d18374524f89696754db0a024c27dc73
The version number targeted for `cluster2` MUST include this PR: https://github.com/cubejs/cluster2/pull/86
|
1.0
|
Move back to upstream cluster2 package if our PR is merged - Update package.json to remove the `git://` url used as an interim solution until the upstream PR is closed, see commit: https://github.com/pelias/api/commit/e3efeb66d18374524f89696754db0a024c27dc73
The version number targeted for `cluster2` MUST include this PR: https://github.com/cubejs/cluster2/pull/86
|
process
|
move back to upstream package if our pr is merged update package json to remove the git url used as an interim solution until the upstream pr is closed see commit the version number targeted for must include this pr
| 1
|
22,001
| 30,504,259,779
|
IssuesEvent
|
2023-07-18 15:45:39
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
init-npm-por-lean-parte_tres 1.0.1 has 5 guarddog issues
|
npm-install-script npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/import-npm/node_modules/init-npm/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/init-npm-por-lean-parte_tres/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/import-npm/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":"\t\tspawn(process.execPath, [path.join(__dirname, 'check.js'), JSON.stringify(this.options)], {\n\t\t\tdetached: true,\n\t\t\tstdio: 'ignore'\n\t\t}).unref();","location":"package/import-npm/node_modules/update-notifier/index.js:97","message":"This package is silently executing another executable"}]}```
|
1.0
|
init-npm-por-lean-parte_tres 1.0.1 has 5 guarddog issues - ```{"npm-install-script":[{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/import-npm/node_modules/init-npm/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/init-npm-por-lean-parte_tres/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/import-npm/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":"\t\tspawn(process.execPath, [path.join(__dirname, 'check.js'), JSON.stringify(this.options)], {\n\t\t\tdetached: true,\n\t\t\tstdio: 'ignore'\n\t\t}).unref();","location":"package/import-npm/node_modules/update-notifier/index.js:97","message":"This package is silently executing another executable"}]}```
|
process
|
init npm por lean parte tres has guarddog issues npm install script npm silent process execution n t t tdetached true n t t tstdio ignore n t t unref location package import npm node modules update notifier index js message this package is silently executing another executable
| 1
|
11,413
| 30,453,357,501
|
IssuesEvent
|
2023-07-16 15:21:42
|
SuperCowPowers/sageworks
|
https://api.github.com/repos/SuperCowPowers/sageworks
|
opened
|
Have a 'refresh now' option for the AWS Service Broker
|
api aws_service_broker application architecture
|
When we 'delete' an artifacts on the AWS Dashboard, there's this weird row deletes, row shows back up and finally row goes away. This is because of the 'eventually consistence' nature of the AWS Service Broker. Anyway lets have the view expose the 'force_refresh' option in some way and have the app call it when a delete happens.
|
1.0
|
Have a 'refresh now' option for the AWS Service Broker - When we 'delete' an artifacts on the AWS Dashboard, there's this weird row deletes, row shows back up and finally row goes away. This is because of the 'eventually consistence' nature of the AWS Service Broker. Anyway lets have the view expose the 'force_refresh' option in some way and have the app call it when a delete happens.
|
non_process
|
have a refresh now option for the aws service broker when we delete an artifacts on the aws dashboard there s this weird row deletes row shows back up and finally row goes away this is because of the eventually consistence nature of the aws service broker anyway lets have the view expose the force refresh option in some way and have the app call it when a delete happens
| 0
|
107,075
| 11,516,690,143
|
IssuesEvent
|
2020-02-14 06:04:38
|
postmanlabs/postman-app-support
|
https://api.github.com/repos/postmanlabs/postman-app-support
|
closed
|
Confusing UX in the `New > Create documentation` flow in the app
|
Microcopy bug product/documentation
|
1. Start with a new Postman installation (signed-in to a free account).
2. Create a collection (with name and description)
3. New > Documentation > (Use collection from this workspace)
Problem 1: The text still says 'Enter the requests you want to document'. This is relevant for the `Create a new API` tab, not the `Use collection from this workspace` tab.
4. Select the collection you created
Problem 2: The name isn't editable, but the description is. Why the discrepancy? The text on the right says 'Enter a title to describe your requests', but I'm not allowed to change the name.
Seen on Postman v6.6, Mac native app
|
1.0
|
Confusing UX in the `New > Create documentation` flow in the app - 1. Start with a new Postman installation (signed-in to a free account).
2. Create a collection (with name and description)
3. New > Documentation > (Use collection from this workspace)
Problem 1: The text still says 'Enter the requests you want to document'. This is relevant for the `Create a new API` tab, not the `Use collection from this workspace` tab.
4. Select the collection you created
Problem 2: The name isn't editable, but the description is. Why the discrepancy? The text on the right says 'Enter a title to describe your requests', but I'm not allowed to change the name.
Seen on Postman v6.6, Mac native app
|
non_process
|
confusing ux in the new create documentation flow in the app start with a new postman installation signed in to a free account create a collection with name and description new documentation use collection from this workspace problem the text still says enter the requests you want to document this is relevant for the create a new api tab not the use collection from this workspace tab select the collection you created problem the name isn t editable but the description is why the discrepancy the text on the right says enter a title to describe your requests but i m not allowed to change the name seen on postman mac native app
| 0
|
12,082
| 14,740,041,711
|
IssuesEvent
|
2021-01-07 08:24:46
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Chattanooga - SA Billing - Late Fee Account List
|
anc-process anp-important ant-bug has attachment
|
In GitLab by @kdjstudios on Oct 3, 2018, 10:58
[Chattanooga.xlsx](/uploads/f892636fe4033036654065e7216841e7/Chattanooga.xlsx)
HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-88683/conversation
|
1.0
|
Chattanooga - SA Billing - Late Fee Account List - In GitLab by @kdjstudios on Oct 3, 2018, 10:58
[Chattanooga.xlsx](/uploads/f892636fe4033036654065e7216841e7/Chattanooga.xlsx)
HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-88683/conversation
|
process
|
chattanooga sa billing late fee account list in gitlab by kdjstudios on oct uploads chattanooga xlsx hd
| 1
|
425,496
| 29,482,405,044
|
IssuesEvent
|
2023-06-02 07:05:05
|
aditya-grover/climate-learn
|
https://api.github.com/repos/aditya-grover/climate-learn
|
opened
|
Statistical Downscaling of other ERA5 Variables
|
documentation
|
So I am attempting to downscale ERA5 Sea Surface Temperature Variable, I was following along your tutorial at NeurIPS2022 CCAI. I noticed there you used 5 degree and 2 Degree Resolutions for 2m_temperature, Why is this done? It is not very clear. For Sea Surface Temperature i have data at 0.25 Degree resolution, but do i need a Coarser resolution to get this code to work for my chosen variable?
|
1.0
|
Statistical Downscaling of other ERA5 Variables - So I am attempting to downscale ERA5 Sea Surface Temperature Variable, I was following along your tutorial at NeurIPS2022 CCAI. I noticed there you used 5 degree and 2 Degree Resolutions for 2m_temperature, Why is this done? It is not very clear. For Sea Surface Temperature i have data at 0.25 Degree resolution, but do i need a Coarser resolution to get this code to work for my chosen variable?
|
non_process
|
statistical downscaling of other variables so i am attempting to downscale sea surface temperature variable i was following along your tutorial at ccai i noticed there you used degree and degree resolutions for temperature why is this done it is not very clear for sea surface temperature i have data at degree resolution but do i need a coarser resolution to get this code to work for my chosen variable
| 0
|
388,841
| 26,784,136,055
|
IssuesEvent
|
2023-02-01 00:27:35
|
automatic-ripping-machine/automatic-ripping-machine
|
https://api.github.com/repos/automatic-ripping-machine/automatic-ripping-machine
|
closed
|
No NVENC hardware encoding option with 1.3.x handbrake installed with the script on Ubuntu 20.04.x
|
documentation Resolved
|
- [ x] I have checked the wiki
- [x ] I have searched for similar issues
**Describe the bug**
So it's no really a bug, just something that I have found.
I have a GTX 1060 NVENC compatible card seen by nvidia-smi tool but handbrake does not list the nvenc encoding option, juste like the hardware encoding is not available with the installed version
**To Reproduce**
Steps to reproduce the behavior:
1. Install an ubuntu 20.04.x vm in promox (pci pasthrought should be enable and the card should be listed when executing the nvida-smi tool)
2. Install ARM using the script
3. use Handbranke-cli to check the encoding options
### Environment
Ubuntu 20.04.5 LTS
ARM: 2.6.10
|
1.0
|
No NVENC hardware encoding option with 1.3.x handbrake installed with the script on Ubuntu 20.04.x - - [ x] I have checked the wiki
- [x ] I have searched for similar issues
**Describe the bug**
So it's no really a bug, just something that I have found.
I have a GTX 1060 NVENC compatible card seen by nvidia-smi tool but handbrake does not list the nvenc encoding option, juste like the hardware encoding is not available with the installed version
**To Reproduce**
Steps to reproduce the behavior:
1. Install an ubuntu 20.04.x vm in promox (pci pasthrought should be enable and the card should be listed when executing the nvida-smi tool)
2. Install ARM using the script
3. use Handbranke-cli to check the encoding options
### Environment
Ubuntu 20.04.5 LTS
ARM: 2.6.10
|
non_process
|
no nvenc hardware encoding option with x handbrake installed with the script on ubuntu x i have checked the wiki i have searched for similar issues describe the bug so it s no really a bug just something that i have found i have a gtx nvenc compatible card seen by nvidia smi tool but handbrake does not list the nvenc encoding option juste like the hardware encoding is not available with the installed version to reproduce steps to reproduce the behavior install an ubuntu x vm in promox pci pasthrought should be enable and the card should be listed when executing the nvida smi tool install arm using the script use handbranke cli to check the encoding options environment ubuntu lts arm
| 0
|
16,280
| 20,904,904,947
|
IssuesEvent
|
2022-03-24 00:19:46
|
quark-engine/quark-engine
|
https://api.github.com/repos/quark-engine/quark-engine
|
closed
|
Update CI for the BladeHawk ruleset release
|
work-in-progress issue-processing-state-06
|
The [release](https://github.com/quark-engine/quark-rules/issues/18) of the BladeHawk ruleset introduced more rules into the quark-rule repo. Since [our smoke test](https://github.com/quark-engine/quark-engine/blob/master/.github/workflows/smoke_test.yml) uses this repo for the accuracy checks, this release has made it inaccurate. Thus, we need to adjust it as well.
|
1.0
|
Update CI for the BladeHawk ruleset release - The [release](https://github.com/quark-engine/quark-rules/issues/18) of the BladeHawk ruleset introduced more rules into the quark-rule repo. Since [our smoke test](https://github.com/quark-engine/quark-engine/blob/master/.github/workflows/smoke_test.yml) uses this repo for the accuracy checks, this release has made it inaccurate. Thus, we need to adjust it as well.
|
process
|
update ci for the bladehawk ruleset release the of the bladehawk ruleset introduced more rules into the quark rule repo since uses this repo for the accuracy checks this release has made it inaccurate thus we need to adjust it as well
| 1
|
19,043
| 25,043,297,496
|
IssuesEvent
|
2022-11-05 00:33:16
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
"Instructions for approvers" text box doesn't seem to do anything
|
product-feedback cba Pri1 azure-devops-pipelines/svc azure-devops-pipelines-process/subsvc
|
When defining an approval on a secured resource (environment, service connection, _et al_) in Azure DevOps Services, there's a "Instructions to approvers (optional)" text box.
This feature doesn't seem to do anything. Text entered in this box is not included in the emails generated by Azure DevOps Services to the approvers, nor is it displayed in the flyout panel on a pipeline run that activates that approval. the box is too small to contain any meaningful amount of text, so it doesn't seem that this intended as a way to document the approval itself.
Is this feature supposed to do anything?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 624da995-5b05-bb05-0afc-c42a1ff5ab87
* Version Independent ID: 3cf8daeb-8d99-845e-dfd8-d0be5f65182c
* Content: [Pipeline deployment approvals - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass)
* Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/approvals.md)
* Service: **azure-devops-pipelines**
* Sub-service: **azure-devops-pipelines-process**
* GitHub Login: @shashban
* Microsoft Alias: **shashban**
|
1.0
|
"Instructions for approvers" text box doesn't seem to do anything -
When defining an approval on a secured resource (environment, service connection, _et al_) in Azure DevOps Services, there's a "Instructions to approvers (optional)" text box.
This feature doesn't seem to do anything. Text entered in this box is not included in the emails generated by Azure DevOps Services to the approvers, nor is it displayed in the flyout panel on a pipeline run that activates that approval. the box is too small to contain any meaningful amount of text, so it doesn't seem that this intended as a way to document the approval itself.
Is this feature supposed to do anything?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 624da995-5b05-bb05-0afc-c42a1ff5ab87
* Version Independent ID: 3cf8daeb-8d99-845e-dfd8-d0be5f65182c
* Content: [Pipeline deployment approvals - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass)
* Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/approvals.md)
* Service: **azure-devops-pipelines**
* Sub-service: **azure-devops-pipelines-process**
* GitHub Login: @shashban
* Microsoft Alias: **shashban**
|
process
|
instructions for approvers text box doesn t seem to do anything when defining an approval on a secured resource environment service connection et al in azure devops services there s a instructions to approvers optional text box this feature doesn t seem to do anything text entered in this box is not included in the emails generated by azure devops services to the approvers nor is it displayed in the flyout panel on a pipeline run that activates that approval the box is too small to contain any meaningful amount of text so it doesn t seem that this intended as a way to document the approval itself is this feature supposed to do anything document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service azure devops pipelines sub service azure devops pipelines process github login shashban microsoft alias shashban
| 1
|
1,119
| 3,592,791,528
|
IssuesEvent
|
2016-02-01 17:15:31
|
DoSomething/quasar
|
https://api.github.com/repos/DoSomething/quasar
|
closed
|
Get Answers from Mobile Commons re: Ingestion
|
#discussion #processing BLOCKER Data 1.0 enhancement small
|
- Is the "First Seen ID" the opt in path or the campaign ID?
- What’s the update on us accessing the database (vs. pinging the API)?
- Can Mobile Commons build something/help with manually subscribing the cohort of people who don’t text in the campaign keyword but express interest in the campaign?
- Is the current version of API the latest version? Are there any changes coming down the pipeline? Is documentation up to date?(edited)
- Can we query the “description” field in a way that is meaningful to us?
|
1.0
|
Get Answers from Mobile Commons re: Ingestion - - Is the "First Seen ID" the opt in path or the campaign ID?
- What’s the update on us accessing the database (vs. pinging the API)?
- Can Mobile Commons build something/help with manually subscribing the cohort of people who don’t text in the campaign keyword but express interest in the campaign?
- Is the current version of API the latest version? Are there any changes coming down the pipeline? Is documentation up to date?(edited)
- Can we query the “description” field in a way that is meaningful to us?
|
process
|
get answers from mobile commons re ingestion is the first seen id the opt in path or the campaign id what’s the update on us accessing the database vs pinging the api can mobile commons build something help with manually subscribing the cohort of people who don’t text in the campaign keyword but express interest in the campaign is the current version of api the latest version are there any changes coming down the pipeline is documentation up to date edited can we query the “description” field in a way that is meaningful to us
| 1
|
190,722
| 15,255,225,880
|
IssuesEvent
|
2021-02-20 15:16:03
|
getpatchwork/patchwork
|
https://api.github.com/repos/getpatchwork/patchwork
|
closed
|
patchwork fails to parse patch with unicode text
|
bug documentation
|
The sourceware instance of patchwork (based on the stable 2.2 branch) fails to parse this patch posted on the list:
https://sourceware.org/pipermail/libc-alpha/2021-January/121972.html
It fails with the following trace:
```
Error when parsing incoming email: OperationalError(1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1")
Traceback (most recent call last):
File "{redacted}/django/db/backends/utils.py", line 84, in _execute
return self.cursor.execute(sql, params)
File "{redacted}/django/db/backends/mysql/base.py", line 71, in execute
return self.cursor.execute(query, args)
File "{redacted}/MySQLdb/cursors.py", line 209, in execute
res = self._query(query)
File "{redacted}/MySQLdb/cursors.py", line 315, in _query
db.query(q)
File "{redacted}/MySQLdb/connections.py", line 239, in query
_mysql.connection.query(self, query)
MySQLdb._exceptions.OperationalError: (1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1")
```
|
1.0
|
patchwork fails to parse patch with unicode text - The sourceware instance of patchwork (based on the stable 2.2 branch) fails to parse this patch posted on the list:
https://sourceware.org/pipermail/libc-alpha/2021-January/121972.html
It fails with the following trace:
```
Error when parsing incoming email: OperationalError(1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1")
Traceback (most recent call last):
File "{redacted}/django/db/backends/utils.py", line 84, in _execute
return self.cursor.execute(sql, params)
File "{redacted}/django/db/backends/mysql/base.py", line 71, in execute
return self.cursor.execute(query, args)
File "{redacted}/MySQLdb/cursors.py", line 209, in execute
res = self._query(query)
File "{redacted}/MySQLdb/cursors.py", line 315, in _query
db.query(q)
File "{redacted}/MySQLdb/connections.py", line 239, in query
_mysql.connection.query(self, query)
MySQLdb._exceptions.OperationalError: (1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1")
```
|
non_process
|
patchwork fails to parse patch with unicode text the sourceware instance of patchwork based on the stable branch fails to parse this patch posted on the list it fails with the following trace error when parsing incoming email operationalerror incorrect string value for column patchwork patchwork patch diff at row traceback most recent call last file redacted django db backends utils py line in execute return self cursor execute sql params file redacted django db backends mysql base py line in execute return self cursor execute query args file redacted mysqldb cursors py line in execute res self query query file redacted mysqldb cursors py line in query db query q file redacted mysqldb connections py line in query mysql connection query self query mysqldb exceptions operationalerror incorrect string value for column patchwork patchwork patch diff at row
| 0
|
225,913
| 24,911,599,670
|
IssuesEvent
|
2022-10-29 23:13:09
|
Piotr1215/dca-prep-kit
|
https://api.github.com/repos/Piotr1215/dca-prep-kit
|
closed
|
core-2.1.0.tgz: 4 vulnerabilities (highest severity is: 7.5) - autoclosed
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>core-2.1.0.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-7753](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-0.0.1.tgz | Transitive | N/A | ❌ |
| [CVE-2022-3517](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | minimatch-3.0.4.tgz | Transitive | N/A | ❌ |
| [CVE-2022-37599](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | loader-utils-2.0.2.tgz | Transitive | N/A | ❌ |
| [CVE-2022-33987](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | got-9.6.0.tgz | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7753</summary>
### Vulnerable Library - <b>trim-0.0.1.tgz</b></p>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/remark-parse/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- mdx-loader-2.1.0.tgz
- mdx-1.6.22.tgz
- remark-parse-8.0.3.tgz
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-10-27</p>
<p>Fix Resolution: trim - 0.0.3</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-3517</summary>
### Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/serve-handler/node_modules/minimatch/package.json,/node_modules/recursive-readdir/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- react-dev-utils-12.0.1.tgz
- recursive-readdir-2.2.2.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-37599</summary>
### Vulnerable Library - <b>loader-utils-2.0.2.tgz</b></p>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- file-loader-6.2.0.tgz
- :x: **loader-utils-2.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js.
<p>Publish Date: 2022-10-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary>
### Vulnerable Library - <b>got-9.6.0.tgz</b></p>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-9.6.0.tgz">https://registry.npmjs.org/got/-/got-9.6.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- update-notifier-5.1.0.tgz
- latest-version-5.1.0.tgz
- package-json-6.5.0.tgz
- :x: **got-9.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
core-2.1.0.tgz: 4 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>core-2.1.0.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-7753](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-0.0.1.tgz | Transitive | N/A | ❌ |
| [CVE-2022-3517](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | minimatch-3.0.4.tgz | Transitive | N/A | ❌ |
| [CVE-2022-37599](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | loader-utils-2.0.2.tgz | Transitive | N/A | ❌ |
| [CVE-2022-33987](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | got-9.6.0.tgz | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7753</summary>
### Vulnerable Library - <b>trim-0.0.1.tgz</b></p>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/remark-parse/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- mdx-loader-2.1.0.tgz
- mdx-1.6.22.tgz
- remark-parse-8.0.3.tgz
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-10-27</p>
<p>Fix Resolution: trim - 0.0.3</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-3517</summary>
### Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/serve-handler/node_modules/minimatch/package.json,/node_modules/recursive-readdir/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- react-dev-utils-12.0.1.tgz
- recursive-readdir-2.2.2.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-37599</summary>
### Vulnerable Library - <b>loader-utils-2.0.2.tgz</b></p>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- file-loader-6.2.0.tgz
- :x: **loader-utils-2.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js.
<p>Publish Date: 2022-10-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary>
### Vulnerable Library - <b>got-9.6.0.tgz</b></p>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-9.6.0.tgz">https://registry.npmjs.org/got/-/got-9.6.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- core-2.1.0.tgz (Root Library)
- update-notifier-5.1.0.tgz
- latest-version-5.1.0.tgz
- package-json-6.5.0.tgz
- :x: **got-9.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
core tgz vulnerabilities highest severity is autoclosed vulnerable library core tgz path to dependency file package json path to vulnerable library node modules loader utils package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high trim tgz transitive n a high minimatch tgz transitive n a high loader utils tgz transitive n a medium got tgz transitive n a details cve vulnerable library trim tgz trim string whitespace library home page a href path to dependency file package json path to vulnerable library node modules remark parse node modules trim package json dependency hierarchy core tgz root library mdx loader tgz mdx tgz remark parse tgz x trim tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package trim are vulnerable to regular expression denial of service redos via trim publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution trim step up your open source security game with mend cve vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules serve handler node modules minimatch package json node modules recursive readdir node modules minimatch package json dependency hierarchy core tgz root library react dev utils tgz recursive readdir tgz x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch step up your open source security game with mend cve vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy core tgz root library file loader tgz x loader utils tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the resourcepath variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend cve vulnerable library got tgz simplified http requests library home page a href path to dependency file package json path to vulnerable library node modules got package json dependency hierarchy core tgz root library update notifier tgz latest version tgz package json tgz x got tgz vulnerable library found in head commit a href found in base branch master vulnerability details the got package before also fixed in for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend
| 0
|
6,109
| 5,285,190,544
|
IssuesEvent
|
2017-02-08 03:34:46
|
OctopusDeploy/Issues
|
https://api.github.com/repos/OctopusDeploy/Issues
|
opened
|
High memory usage when applying deployment manifest retention policy
|
area/scale feature/performance
|
As reported here: http://help.octopusdeploy.com/discussions/problems/51077
Jeff's (the reporting customer) diagnosis seems correct. We are loading _all_ variable sets for all completed and failed deployments. We should stream these.
|
True
|
High memory usage when applying deployment manifest retention policy - As reported here: http://help.octopusdeploy.com/discussions/problems/51077
Jeff's (the reporting customer) diagnosis seems correct. We are loading _all_ variable sets for all completed and failed deployments. We should stream these.
|
non_process
|
high memory usage when applying deployment manifest retention policy as reported here jeff s the reporting customer diagnosis seems correct we are loading all variable sets for all completed and failed deployments we should stream these
| 0
|
89,654
| 25,867,674,376
|
IssuesEvent
|
2022-12-13 22:30:57
|
USGCRP/gcis
|
https://api.github.com/repos/USGCRP/gcis
|
closed
|
Handle special characters
|
help wanted type content type technical priority medium on hold for rebuild
|
Example 1 : When an organization with special character(s) is added as an org title, the automatically generated identifier ignores the special characters.
Example 2: Person name with special character(s) is currently hard to find using the search query unless the last name or first name doesn't have special characters in it.
We should either teach the system to handle special characters or ignore using them to avoid redundancy.
|
1.0
|
Handle special characters - Example 1 : When an organization with special character(s) is added as an org title, the automatically generated identifier ignores the special characters.
Example 2: Person name with special character(s) is currently hard to find using the search query unless the last name or first name doesn't have special characters in it.
We should either teach the system to handle special characters or ignore using them to avoid redundancy.
|
non_process
|
handle special characters example when an organization with special character s is added as an org title the automatically generated identifier ignores the special characters example person name with special character s is currently hard to find using the search query unless the last name or first name doesn t have special characters in it we should either teach the system to handle special characters or ignore using them to avoid redundancy
| 0
|
9,397
| 2,615,147,568
|
IssuesEvent
|
2015-03-01 06:23:53
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Add case studies as new resources on front page
|
auto-migrated Milestone-3 Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `v...@google.com` on 30 Sep 2010 at 8:33
|
1.0
|
Add case studies as new resources on front page - ```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `v...@google.com` on 30 Sep 2010 at 8:33
|
non_process
|
add case studies as new resources on front page what steps will reproduce the problem what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by v google com on sep at
| 0
|
315,209
| 9,607,783,061
|
IssuesEvent
|
2019-05-11 22:16:19
|
x-klanas/Wrath
|
https://api.github.com/repos/x-klanas/Wrath
|
opened
|
SteamVR actions
|
2 points high priority user story
|
As a player I want to be able to customize the controller bindings.
- [ ] the required controller bindings must be created and setup appropriately
- [ ] the default bindings for as many controllers as possible must be set
|
1.0
|
SteamVR actions - As a player I want to be able to customize the controller bindings.
- [ ] the required controller bindings must be created and setup appropriately
- [ ] the default bindings for as many controllers as possible must be set
|
non_process
|
steamvr actions as a player i want to be able to customize the controller bindings the required controller bindings must be created and setup appropriately the default bindings for as many controllers as possible must be set
| 0
|
141,201
| 11,403,655,731
|
IssuesEvent
|
2020-01-31 07:50:56
|
NethermindEth/nethermind
|
https://api.github.com/repos/NethermindEth/nethermind
|
closed
|
Not all ethereum tests running on MacOs and Linux
|
devops test
|


I have these failing tests on one of the branches (proofs). It fails on Windows build but passed on Linux and MacOS. After short investigation I can see that Linux and MacOS simply do not run some of the tests.
|
1.0
|
Not all ethereum tests running on MacOs and Linux - 

I have these failing tests on one of the branches (proofs). It fails on Windows build but passed on Linux and MacOS. After short investigation I can see that Linux and MacOS simply do not run some of the tests.
|
non_process
|
not all ethereum tests running on macos and linux i have these failing tests on one of the branches proofs it fails on windows build but passed on linux and macos after short investigation i can see that linux and macos simply do not run some of the tests
| 0
|
11,951
| 14,713,191,838
|
IssuesEvent
|
2021-01-05 10:00:47
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
unable to run migrations in an empty PostGIS db
|
bug/2-confirmed kind/bug process/candidate team/migrations
|
<!--
Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client.
Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports
-->
## Bug description
Copied from https://github.com/prisma/prisma/discussions/4732
<!-- A clear and concise description of what the bug is. -->
I'm struggling to use the migration tool after I upgraded to 2.13.0.
I have a few spatial columns that I manage with scripts besides the Prisma migration tool.
My pipeline used to be this:
1. apply Prisma migrations
2. run manual migrations and add / alter spatial columns if needed
but with 2.13.0, I can no longer run any Prisma migration on a fresh PostGIS database because of this error:
```
Error: P3005
The database schema for `localhost:5432` is not empty. Read more about how to baseline an existing production database: https://pris.ly/d/migrate-baseline
```
Enabling PostGIS creates a `spatial_ref_sys` table, so technically the database is not empty, but it should not affect migrations.
I know Prisma does not yet support the PostGIS extension, but I'm looking for a workaround.
I'm also a Nexus and nexus-plugin-prisma user which doesn't support 2.12.x, so I can't downgrade reliably.
## How to reproduce
Steps to reproduce the behavior:
1. Have an empty PostGIS database (I used https://hub.docker.com/r/postgis/postgis)
1. Run `prisma migrate dev --preview-feature`
1. See error
## Expected behavior
Migrations run without an error.
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: PopOS
- Database: PostgreSQL with PostGIS
- Node.js version: v14.13.0
- Prisma version:
```
@prisma/cli : 2.13.0
@prisma/client : 2.13.0
Current platform : debian-openssl-1.1.x
Query Engine : query-engine 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x)
Migration Engine : migration-engine-cli 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x)
Introspection Engine : introspection-core 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x)
Format Binary : prisma-fmt 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x)
Studio : 0.329.0
```
|
1.0
|
unable to run migrations in an empty PostGIS db - <!--
Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client.
Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports
-->
## Bug description
Copied from https://github.com/prisma/prisma/discussions/4732
<!-- A clear and concise description of what the bug is. -->
I'm struggling to use the migration tool after I upgraded to 2.13.0.
I have a few spatial columns that I manage with scripts besides the Prisma migration tool.
My pipeline used to be this:
1. apply Prisma migrations
2. run manual migrations and add / alter spatial columns if needed
but with 2.13.0, I can no longer run any Prisma migration on a fresh PostGIS database because of this error:
```
Error: P3005
The database schema for `localhost:5432` is not empty. Read more about how to baseline an existing production database: https://pris.ly/d/migrate-baseline
```
Enabling PostGIS creates a `spatial_ref_sys` table, so technically the database is not empty, but it should not affect migrations.
I know Prisma does not yet support the PostGIS extension, but I'm looking for a workaround.
I'm also a Nexus and nexus-plugin-prisma user which doesn't support 2.12.x, so I can't downgrade reliably.
## How to reproduce
Steps to reproduce the behavior:
1. Have an empty PostGIS database (I used https://hub.docker.com/r/postgis/postgis)
1. Run `prisma migrate dev --preview-feature`
1. See error
## Expected behavior
Migrations run without an error.
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: PopOS
- Database: PostgreSQL with PostGIS
- Node.js version: v14.13.0
- Prisma version:
```
@prisma/cli : 2.13.0
@prisma/client : 2.13.0
Current platform : debian-openssl-1.1.x
Query Engine : query-engine 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x)
Migration Engine : migration-engine-cli 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x)
Introspection Engine : introspection-core 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x)
Format Binary : prisma-fmt 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x)
Studio : 0.329.0
```
|
process
|
unable to run migrations in an empty postgis db thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description copied from i m struggling to use the migration tool after i upgraded to i have a few spatial columns that i manage with scripts besides the prisma migration tool my pipeline used to be this apply prisma migrations run manual migrations and add alter spatial columns if needed but with i can no longer run any prisma migration on a fresh postgis database because of this error error the database schema for localhost is not empty read more about how to baseline an existing production database enabling postgis creates a spatial ref sys table so technically the database is not empty but it should not affect migrations i know prisma does not yet support the postgis extension but i m looking for a workaround i m also a nexus and nexus plugin prisma user which doesn t support x so i can t downgrade reliably how to reproduce steps to reproduce the behavior have an empty postgis database i used run prisma migrate dev preview feature see error expected behavior migrations run without an error environment setup os popos database postgresql with postgis node js version prisma version prisma cli prisma client current platform debian openssl x query engine query engine at node modules prisma engines query engine debian openssl x migration engine migration engine cli at node modules prisma engines migration engine debian openssl x introspection engine introspection core at node modules prisma engines introspection engine debian openssl x format binary prisma fmt at node modules prisma engines prisma fmt debian openssl x studio
| 1
|
204,205
| 15,428,793,361
|
IssuesEvent
|
2021-03-06 00:51:31
|
comit-network/xmr-btc-swap
|
https://api.github.com/repos/comit-network/xmr-btc-swap
|
opened
|
Alice has problems redeeming
|
bug test/stagenet demo
|
Scenario:
1. Bob locks BTC
2. Alice locks XMR
3. Alice runs into Electrum error when redeeming
4. Bob runs into Electrum error when watching for redeem
Alice's error:
```
Mar 05 12:50:46.784 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:48.689 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:50.493 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:52.302 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:54.102 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:55.907 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
```
On Alice's (ASB) side this is a severe problem where we print to the logs about every half second spamming the logs really fast. Eventually the ASB restarts (at the moment I am not sure what causes the restart, but it could be systemd picking up the log load?).
Bob's error log:
```
Monero lock tx has 10 out of 10 confirmations txid=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0
Error: Transient errors should be retried
Caused by:
Electrum client error
```
Lock transaction hashes where this problem was recorded (Bob):
* https://blockstream.info/testnet/tx/3e305e6895211f974718d3ad83e0649d3a090a726f3773c9dc415fc9888a9e24
* https://monero-stagenet.exan.tech/search?value=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0
Investigation needed why Alice runs into this error - is it load related (i.e. we spam the public Electrum instance too much) or is it related to other problems (e.g. specific addresses, ...).
|
1.0
|
Alice has problems redeeming - Scenario:
1. Bob locks BTC
2. Alice locks XMR
3. Alice runs into Electrum error when redeeming
4. Bob runs into Electrum error when watching for redeem
Alice's error:
```
Mar 05 12:50:46.784 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:48.689 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:50.493 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:52.302 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:54.102 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
Mar 05 12:50:55.907 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying...
```
On Alice's (ASB) side this is a severe problem where we print to the logs about every half second spamming the logs really fast. Eventually the ASB restarts (at the moment I am not sure what causes the restart, but it could be systemd picking up the log load?).
Bob's error log:
```
Monero lock tx has 10 out of 10 confirmations txid=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0
Error: Transient errors should be retried
Caused by:
Electrum client error
```
Lock transaction hashes where this problem was recorded (Bob):
* https://blockstream.info/testnet/tx/3e305e6895211f974718d3ad83e0649d3a090a726f3773c9dc415fc9888a9e24
* https://monero-stagenet.exan.tech/search?value=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0
Investigation needed why Alice runs into this error - is it load related (i.e. we spam the public Electrum instance too much) or is it related to other problems (e.g. specific addresses, ...).
|
non_process
|
alice has problems redeeming scenario bob locks btc alice locks xmr alice runs into electrum error when redeeming bob runs into electrum error when watching for redeem alice s error mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying on alice s asb side this is a severe problem where we print to the logs about every half second spamming the logs really fast eventually the asb restarts at the moment i am not sure what causes the restart but it could be systemd picking up the log load bob s error log monero lock tx has out of confirmations txid error transient errors should be retried caused by electrum client error lock transaction hashes where this problem was recorded bob investigation needed why alice runs into this error is it load related i e we spam the public electrum instance too much or is it related to other problems e g specific addresses
| 0
|
4,247
| 7,187,154,477
|
IssuesEvent
|
2018-02-02 03:16:18
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
Monitor speed improvement
|
monitors-all status-inprocess type-enhancement
|
For non-contract accounts the only thing one needs to check to see if the block is of interest is if the account's balance changed.
The only way for a non-contract account to have been involved in a transaction is if it
a. sent a transaction in which case its balance changed via gas usage,
b. received ether in which case its balance changed, or
c. was involved in two, or four, or six, or 2*n transactions that exactly offset it in the same block in which case its balance will not have changed.
This last can be found in --deep scan, but we might be able to shortcut the search if the account is not a smart contract and it's balance did not change.
Question: is querying for balance worse than or faster than scanning the transactions? My guess is the former by a very large margin.
|
1.0
|
Monitor speed improvement - For non-contract accounts the only thing one needs to check to see if the block is of interest is if the account's balance changed.
The only way for a non-contract account to have been involved in a transaction is if it
a. sent a transaction in which case its balance changed via gas usage,
b. received ether in which case its balance changed, or
c. was involved in two, or four, or six, or 2*n transactions that exactly offset it in the same block in which case its balance will not have changed.
This last can be found in --deep scan, but we might be able to shortcut the search if the account is not a smart contract and it's balance did not change.
Question: is querying for balance worse than or faster than scanning the transactions? My guess is the former by a very large margin.
|
process
|
monitor speed improvement for non contract accounts the only thing one needs to check to see if the block is of interest is if the account s balance changed the only way for a non contract account to have been involved in a transaction is if it a sent a transaction in which case its balance changed via gas usage b received ether in which case its balance changed or c was involved in two or four or six or n transactions that exactly offset it in the same block in which case its balance will not have changed this last can be found in deep scan but we might be able to shortcut the search if the account is not a smart contract and it s balance did not change question is querying for balance worse than or faster than scanning the transactions my guess is the former by a very large margin
| 1
|
6,458
| 9,546,567,737
|
IssuesEvent
|
2019-05-01 20:19:41
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Department of State: Transcripts
|
Apply Process Approved Requirements Ready State Dept.
|
Who: Student Applicant
What: Transcript page
Why: The student is required to submit a transcript with their application.
- There will be a header "Transcripts"
- There will be a list of transcripts the user has stored in their USAJOBS profile
- The user can select the transcript they would like to submit with their application
- The "Upload transcript" link will take the user to their USAJOBS profile to upload a transcript (this will open in a new window)
- When the user returns to their application there will be a "refresh transcripts" button to update their application (this was discussed in our sprint planning meeting- because the transcripts are uploaded in USAJOBS we need to pull the new transcript into the application)
- When the user clicks the "view" button it will act the same as it does on USAJOBS (this was discussed during sprint planning- the view functioanlity will be dependent on the users browser)
- The "Back" Button will take the user to the "Experience & References" page
- The "Save and continue" button will take the user to the "Languages & Skills" page.
- New Transcripts are added and saved on USAJOBS.
Public Link: https://opm.invisionapp.com/share/ZEPNZR09Q54
|
1.0
|
Department of State: Transcripts - Who: Student Applicant
What: Transcript page
Why: The student is required to submit a transcript with their application.
- There will be a header "Transcripts"
- There will be a list of transcripts the user has stored in their USAJOBS profile
- The user can select the transcript they would like to submit with their application
- The "Upload transcript" link will take the user to their USAJOBS profile to upload a transcript (this will open in a new window)
- When the user returns to their application there will be a "refresh transcripts" button to update their application (this was discussed in our sprint planning meeting- because the transcripts are uploaded in USAJOBS we need to pull the new transcript into the application)
- When the user clicks the "view" button it will act the same as it does on USAJOBS (this was discussed during sprint planning- the view functioanlity will be dependent on the users browser)
- The "Back" Button will take the user to the "Experience & References" page
- The "Save and continue" button will take the user to the "Languages & Skills" page.
- New Transcripts are added and saved on USAJOBS.
Public Link: https://opm.invisionapp.com/share/ZEPNZR09Q54
|
process
|
department of state transcripts who student applicant what transcript page why the student is required to submit a transcript with their application there will be a header transcripts there will be a list of transcripts the user has stored in their usajobs profile the user can select the transcript they would like to submit with their application the upload transcript link will take the user to their usajobs profile to upload a transcript this will open in a new window when the user returns to their application there will be a refresh transcripts button to update their application this was discussed in our sprint planning meeting because the transcripts are uploaded in usajobs we need to pull the new transcript into the application when the user clicks the view button it will act the same as it does on usajobs this was discussed during sprint planning the view functioanlity will be dependent on the users browser the back button will take the user to the experience references page the save and continue button will take the user to the languages skills page new transcripts are added and saved on usajobs public link
| 1
|
39,369
| 9,416,542,584
|
IssuesEvent
|
2019-04-10 14:52:43
|
SasView/sasview
|
https://api.github.com/repos/SasView/sasview
|
opened
|
5.0 plugin product model error ?
|
defect
|
In my local build of 5.0, ESS_GUI, I made a plugin model for sphere times hardsphere
This does not work - likely a feature not yet coded - see error message below.
ALSO need to decide how we handle the beta(Q) options, structure_factor_mode, radius_effective_mode etc.
15:43:17 - INFO: 2019-04-10 15:43:17
15:43:17 - ERROR: Traceback (most recent call last):
File "C:\sasview42\sasview\src\sas\sascalc\data_util\calcthread.py", line 274, in _run
self.compute(*args, **kwargs)
File "C:\sasview42\sasview\src\sas\qtgui\Perspectives\Fitting\ModelThread.py", line 211, in compute
intermediate_results = intermediate_results()
TypeError: 'list' object is not callable
|
1.0
|
5.0 plugin product model error ? - In my local build of 5.0, ESS_GUI, I made a plugin model for sphere times hardsphere
This does not work - likely a feature not yet coded - see error message below.
ALSO need to decide how we handle the beta(Q) options, structure_factor_mode, radius_effective_mode etc.
15:43:17 - INFO: 2019-04-10 15:43:17
15:43:17 - ERROR: Traceback (most recent call last):
File "C:\sasview42\sasview\src\sas\sascalc\data_util\calcthread.py", line 274, in _run
self.compute(*args, **kwargs)
File "C:\sasview42\sasview\src\sas\qtgui\Perspectives\Fitting\ModelThread.py", line 211, in compute
intermediate_results = intermediate_results()
TypeError: 'list' object is not callable
|
non_process
|
plugin product model error in my local build of ess gui i made a plugin model for sphere times hardsphere this does not work likely a feature not yet coded see error message below also need to decide how we handle the beta q options structure factor mode radius effective mode etc info error traceback most recent call last file c sasview src sas sascalc data util calcthread py line in run self compute args kwargs file c sasview src sas qtgui perspectives fitting modelthread py line in compute intermediate results intermediate results typeerror list object is not callable
| 0
|
6,502
| 9,576,150,108
|
IssuesEvent
|
2019-05-07 08:25:09
|
DCRGraphsNet/DCROpenCaseManager
|
https://api.github.com/repos/DCRGraphsNet/DCROpenCaseManager
|
closed
|
As a social worker I should be able to see the status of single processes of a child. (3)
|
Must have Page: Child Page: Process Sprint5
|
What is the status actually (ask Syddjurs)?
Status of each process of a child.
How far in the process are we?
Show Green (all ok) or Red (deadline broken)
|
1.0
|
As a social worker I should be able to see the status of single processes of a child. (3) - What is the status actually (ask Syddjurs)?
Status of each process of a child.
How far in the process are we?
Show Green (all ok) or Red (deadline broken)
|
process
|
as a social worker i should be able to see the status of single processes of a child what is the status actually ask syddjurs status of each process of a child how far in the process are we show green all ok or red deadline broken
| 1
|
93,355
| 26,933,048,691
|
IssuesEvent
|
2023-02-07 18:15:36
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Add support for VS2022 build chain
|
module: build module: windows triaged
|
**Requirements**
- PyTorch can be compiled with VS2022 build chain.
**Acceptance Criteria**
- CD pipelines (`ciflow-binaries`) are producing binaries using VS 2022.
cc @malfet @seemethere @peterjc123 @mszhanyi @skyline75489 @nbcsm
|
1.0
|
Add support for VS2022 build chain - **Requirements**
- PyTorch can be compiled with VS2022 build chain.
**Acceptance Criteria**
- CD pipelines (`ciflow-binaries`) are producing binaries using VS 2022.
cc @malfet @seemethere @peterjc123 @mszhanyi @skyline75489 @nbcsm
|
non_process
|
add support for build chain requirements pytorch can be compiled with build chain acceptance criteria cd pipelines ciflow binaries are producing binaries using vs cc malfet seemethere mszhanyi nbcsm
| 0
|
14,143
| 17,034,746,346
|
IssuesEvent
|
2021-07-05 04:51:34
|
theislab/scanpy
|
https://api.github.com/repos/theislab/scanpy
|
opened
|
Better handling of upstream releases
|
Development Process 🚀 Enhancement ✨
|
It would be nice if we had a better way of handling upstream releases. E.g. when pandas makes a release it would be good that we had tested against their release candidates, or if we had a good process for dealing with bugs if they do occur.
One think we could do, is defensively pin dependencies to below their current release series. I don't like doing this since I think it's pretty restrictive when most of the time we don't have issues. Maybe we could do this for breaking releases, but that wouldn't have prevented issues like #1917.
It would be nice to automate the process of testing against upstream release candidates. Basically, when something comes out, we build against it so we can report issues early and don't have to deal with it in live releases. I'm not sure how to do this with `pip search` not working anymore.
|
1.0
|
Better handling of upstream releases - It would be nice if we had a better way of handling upstream releases. E.g. when pandas makes a release it would be good that we had tested against their release candidates, or if we had a good process for dealing with bugs if they do occur.
One think we could do, is defensively pin dependencies to below their current release series. I don't like doing this since I think it's pretty restrictive when most of the time we don't have issues. Maybe we could do this for breaking releases, but that wouldn't have prevented issues like #1917.
It would be nice to automate the process of testing against upstream release candidates. Basically, when something comes out, we build against it so we can report issues early and don't have to deal with it in live releases. I'm not sure how to do this with `pip search` not working anymore.
|
process
|
better handling of upstream releases it would be nice if we had a better way of handling upstream releases e g when pandas makes a release it would be good that we had tested against their release candidates or if we had a good process for dealing with bugs if they do occur one think we could do is defensively pin dependencies to below their current release series i don t like doing this since i think it s pretty restrictive when most of the time we don t have issues maybe we could do this for breaking releases but that wouldn t have prevented issues like it would be nice to automate the process of testing against upstream release candidates basically when something comes out we build against it so we can report issues early and don t have to deal with it in live releases i m not sure how to do this with pip search not working anymore
| 1
|
1,709
| 3,897,596,180
|
IssuesEvent
|
2016-04-16 14:41:22
|
justarrived/just_match_api
|
https://api.github.com/repos/justarrived/just_match_api
|
closed
|
Frilans Finans invoice implementation
|
3rd-party-service in progress must have
|
Currently the external API is not live yet and too unstable to mock.
|
1.0
|
Frilans Finans invoice implementation - Currently the external API is not live yet and too unstable to mock.
|
non_process
|
frilans finans invoice implementation currently the external api is not live yet and too unstable to mock
| 0
|
412,726
| 27,869,434,476
|
IssuesEvent
|
2023-03-21 12:33:51
|
s1dlx/sd-webui-bayesian-merger
|
https://api.github.com/repos/s1dlx/sd-webui-bayesian-merger
|
closed
|
Need a brief description for --scorer_model_dir
|
documentation
|
First off Awesome work! Its lacking a quick sentence about how it deals with the scorer models.
Not sure if it tries all models or can even load more than one.
I copied all the models from the https://github.com/Xerxemi/sdweb-auto-MBW plugin. Are there others that work with it is that even a good idea? Im assuming this option is not needed if I just drop models into that folder??
Thanks for all your hard work!
|
1.0
|
Need a brief description for --scorer_model_dir - First off Awesome work! Its lacking a quick sentence about how it deals with the scorer models.
Not sure if it tries all models or can even load more than one.
I copied all the models from the https://github.com/Xerxemi/sdweb-auto-MBW plugin. Are there others that work with it is that even a good idea? Im assuming this option is not needed if I just drop models into that folder??
Thanks for all your hard work!
|
non_process
|
need a brief description for scorer model dir first off awesome work its lacking a quick sentence about how it deals with the scorer models not sure if it tries all models or can even load more than one i copied all the models from the plugin are there others that work with it is that even a good idea im assuming this option is not needed if i just drop models into that folder thanks for all your hard work
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.