Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
113,244
17,116,224,153
IssuesEvent
2021-07-11 12:10:03
theHinneh/ha
https://api.github.com/repos/theHinneh/ha
closed
CVE-2019-1010266 (Medium) detected in lodash-3.10.1.tgz
security vulnerability
## CVE-2019-1010266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: ha/backend/package.json</p> <p>Path to vulnerable library: ha/backend/node_modules/ioredis/node_modules/lodash/package.json,ha/backend/node_modules/kafka-node/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - mosca-2.8.3.tgz (Root Library) - ioredis-1.15.1.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/theHinneh/ha/commit/b67d33dd9df9e05b70466e310843976220230240">b67d33dd9df9e05b70466e310843976220230240</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2019-07-17</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-1010266 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2019-1010266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: ha/backend/package.json</p> <p>Path to vulnerable library: ha/backend/node_modules/ioredis/node_modules/lodash/package.json,ha/backend/node_modules/kafka-node/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - mosca-2.8.3.tgz (Root Library) - ioredis-1.15.1.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/theHinneh/ha/commit/b67d33dd9df9e05b70466e310843976220230240">b67d33dd9df9e05b70466e310843976220230240</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2019-07-17</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file ha backend package json path to vulnerable library ha backend node modules ioredis node modules lodash package json ha backend node modules kafka node node modules lodash package json dependency hierarchy mosca tgz root library ioredis tgz x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
328,578
28,125,935,891
IssuesEvent
2023-03-31 17:42:55
MicrosoftDocs/visualstudio-docs
https://api.github.com/repos/MicrosoftDocs/visualstudio-docs
closed
How to target an Azure cloud PC / Windows365 that I have access to
doc-bug visual-studio-windows/prod vs-ide-test/tech Pri2
How to target an Azure cloud PC / Windows365 that I have access to via Remote Desktop app. I don't see ways to target the machine, as I have no clue what would be the public name, as it seems to be behind a RDP gateway. Is there ways to proxy the connection via that? --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 90ec9455-ef64-8854-94bc-9be28887a6a4 * Version Independent ID: 9adeea5e-7ba1-065f-5de0-aae9147f9c85 * Content: [Remote Testing in Visual Studio - Visual Studio (Windows)](https://learn.microsoft.com/en-us/visualstudio/test/remote-testing?view=vs-2022) * Content Source: [docs/test/remote-testing.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/main/docs/test/remote-testing.md) * Product: **visual-studio-windows** * Technology: **vs-ide-test** * GitHub Login: @Mikejo5000 * Microsoft Alias: **mikejo**
1.0
How to target an Azure cloud PC / Windows365 that I have access to - How to target an Azure cloud PC / Windows365 that I have access to via Remote Desktop app. I don't see ways to target the machine, as I have no clue what would be the public name, as it seems to be behind a RDP gateway. Is there ways to proxy the connection via that? --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 90ec9455-ef64-8854-94bc-9be28887a6a4 * Version Independent ID: 9adeea5e-7ba1-065f-5de0-aae9147f9c85 * Content: [Remote Testing in Visual Studio - Visual Studio (Windows)](https://learn.microsoft.com/en-us/visualstudio/test/remote-testing?view=vs-2022) * Content Source: [docs/test/remote-testing.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/main/docs/test/remote-testing.md) * Product: **visual-studio-windows** * Technology: **vs-ide-test** * GitHub Login: @Mikejo5000 * Microsoft Alias: **mikejo**
non_process
how to target an azure cloud pc that i have access to how to target an azure cloud pc that i have access to via remote desktop app i don t see ways to target the machine as i have no clue what would be the public name as it seems to be behind a rdp gateway is there ways to proxy the connection via that document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product visual studio windows technology vs ide test github login microsoft alias mikejo
0
178,296
13,772,140,416
IssuesEvent
2020-10-07 23:41:53
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Schedule is not marked as a mandatory field in a CronJob create page
[zube]: To Test kind/bug
**On master-head - commit id: `33f755f18`** - Workloads --> Cron Job --> create - Schedule is not marked as a mandatory field in a CronJob create page <img width="950" alt="Screen Shot 2020-09-17 at 3 03 21 PM" src="https://user-images.githubusercontent.com/26032343/93533061-0dc75800-f8f7-11ea-9d0c-0a23c18838da.png"> **Expected Result:** Cron Job should be marked as a mandatory field
1.0
Schedule is not marked as a mandatory field in a CronJob create page - **On master-head - commit id: `33f755f18`** - Workloads --> Cron Job --> create - Schedule is not marked as a mandatory field in a CronJob create page <img width="950" alt="Screen Shot 2020-09-17 at 3 03 21 PM" src="https://user-images.githubusercontent.com/26032343/93533061-0dc75800-f8f7-11ea-9d0c-0a23c18838da.png"> **Expected Result:** Cron Job should be marked as a mandatory field
non_process
schedule is not marked as a mandatory field in a cronjob create page on master head commit id workloads cron job create schedule is not marked as a mandatory field in a cronjob create page img width alt screen shot at pm src expected result cron job should be marked as a mandatory field
0
13,513
3,343,449,948
IssuesEvent
2015-11-15 14:24:26
bolt/bolt
https://api.github.com/repos/bolt/bolt
closed
Multiple relation fields don't clear when single relation used
Blocking release Bug Needs Acceptance Test Needs Unit Test Regression
The value(s) fetched in `Bolt\Storage\Field\Type\RelationType::persist()` become `EntityProxy`
2.0
Multiple relation fields don't clear when single relation used - The value(s) fetched in `Bolt\Storage\Field\Type\RelationType::persist()` become `EntityProxy`
non_process
multiple relation fields don t clear when single relation used the value s fetched in bolt storage field type relationtype persist become entityproxy
0
93,588
3,906,046,838
IssuesEvent
2016-04-19 07:04:24
Captianrock/android_PV
https://api.github.com/repos/Captianrock/android_PV
opened
Dynamic updates for apps with traces
High Priority
Dynamically update the list of apps that have been analyzed by the user.
1.0
Dynamic updates for apps with traces - Dynamically update the list of apps that have been analyzed by the user.
non_process
dynamic updates for apps with traces dynamically update the list of apps that have been analyzed by the user
0
733,295
25,299,478,315
IssuesEvent
2022-11-17 09:38:06
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
closed
[Model Serving]: Support Model Creation in Global View
kind/enhancement priority/blocker feature/model-serving
### Feature description Follow up #648 Add the ability to deploy a model in the global view ([Mocks](https://www.sketch.com/s/113593f8-5970-49d6-a352-709b07639127/a/EL8AWlg)) ### Describe alternatives you've considered _No response_ ### Anything else? _No response_
1.0
[Model Serving]: Support Model Creation in Global View - ### Feature description Follow up #648 Add the ability to deploy a model in the global view ([Mocks](https://www.sketch.com/s/113593f8-5970-49d6-a352-709b07639127/a/EL8AWlg)) ### Describe alternatives you've considered _No response_ ### Anything else? _No response_
non_process
support model creation in global view feature description follow up add the ability to deploy a model in the global view describe alternatives you ve considered no response anything else no response
0
16,661
21,728,177,662
IssuesEvent
2022-05-11 09:32:33
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Extract RestClient creation logic out of the ElasticsearchClient
kind/toil team/process-automation area/maintainability
**Description** The `ElasticsearchClient` class creates a new `RestClient` based on user configuration passed via `ElasticsearchExporterConfiguration`. Splitting this out will allow us to test that the rest clients are properly constructed, and also allow us to reuse the same logic to construct high level REST clients for testing.
1.0
Extract RestClient creation logic out of the ElasticsearchClient - **Description** The `ElasticsearchClient` class creates a new `RestClient` based on user configuration passed via `ElasticsearchExporterConfiguration`. Splitting this out will allow us to test that the rest clients are properly constructed, and also allow us to reuse the same logic to construct high level REST clients for testing.
process
extract restclient creation logic out of the elasticsearchclient description the elasticsearchclient class creates a new restclient based on user configuration passed via elasticsearchexporterconfiguration splitting this out will allow us to test that the rest clients are properly constructed and also allow us to reuse the same logic to construct high level rest clients for testing
1
4,720
7,552,846,759
IssuesEvent
2018-04-19 02:43:42
UnbFeelings/unb-feelings-docs
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
closed
[Não Conformidade] Relatório Final não existe
Processo Qualidade invalid
@UnbFeelings/process Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o) fora auditada o artefato Relatório Final, resultante da atividade [Relatório Final de Medição](https://github.com/UnbFeelings/unb-feelings-docs/wiki/Processo#317-atividade-relatório-final-de-medição). ### Descrição Foi identificado que não foram realizadas coletas de métricas de código fonte, ou se foram, não estão descritas em um artefato de acordo com o proposto pelo processo. #### Recomendações É recomendado a integração de ferramentas de análise de código para que as métricas possam ser geradas automaticamente, sem que haja a necessidade de atribuir esta tarefa a uma pessoa. No entanto recomenda-se definir um responsável pela elaboração do relatório de métricas. #### Detalhes **Auditor**: Jonathan Rufino **Técnica de Audição**: Checklist **Tipo:** Medição e Análise **Prazo:** 23/04/2018
1.0
[Não Conformidade] Relatório Final não existe - @UnbFeelings/process Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o) fora auditada o artefato Relatório Final, resultante da atividade [Relatório Final de Medição](https://github.com/UnbFeelings/unb-feelings-docs/wiki/Processo#317-atividade-relatório-final-de-medição). ### Descrição Foi identificado que não foram realizadas coletas de métricas de código fonte, ou se foram, não estão descritas em um artefato de acordo com o proposto pelo processo. #### Recomendações É recomendado a integração de ferramentas de análise de código para que as métricas possam ser geradas automaticamente, sem que haja a necessidade de atribuir esta tarefa a uma pessoa. No entanto recomenda-se definir um responsável pela elaboração do relatório de métricas. #### Detalhes **Auditor**: Jonathan Rufino **Técnica de Audição**: Checklist **Tipo:** Medição e Análise **Prazo:** 23/04/2018
process
relatório final não existe unbfeelings process perante critérios definidos para as fora auditada o artefato relatório final resultante da atividade descrição foi identificado que não foram realizadas coletas de métricas de código fonte ou se foram não estão descritas em um artefato de acordo com o proposto pelo processo recomendações é recomendado a integração de ferramentas de análise de código para que as métricas possam ser geradas automaticamente sem que haja a necessidade de atribuir esta tarefa a uma pessoa no entanto recomenda se definir um responsável pela elaboração do relatório de métricas detalhes auditor jonathan rufino técnica de audição checklist tipo medição e análise prazo
1
3,990
6,918,318,910
IssuesEvent
2017-11-29 11:43:50
nlbdev/pipeline
https://api.github.com/repos/nlbdev/pipeline
closed
Acrynyms with genitive "s" / Akronymer med genitivs-s
enhancement pre-processing Priority:1 - Low
(norwegian) *from Trello-board (@matskober):* Legge inn akronymer med genitivs-s (f.eks. NLBs. I punktskrift skal det være et tegn som markerer skillet mellom akronymet og genitivs-s - 56). Liste over akronymer er skaffet fra Mari
1.0
Acrynyms with genitive "s" / Akronymer med genitivs-s - (norwegian) *from Trello-board (@matskober):* Legge inn akronymer med genitivs-s (f.eks. NLBs. I punktskrift skal det være et tegn som markerer skillet mellom akronymet og genitivs-s - 56). Liste over akronymer er skaffet fra Mari
process
acrynyms with genitive s akronymer med genitivs s norwegian from trello board matskober legge inn akronymer med genitivs s f eks nlbs i punktskrift skal det være et tegn som markerer skillet mellom akronymet og genitivs s liste over akronymer er skaffet fra mari
1
13,650
8,306,928,655
IssuesEvent
2018-09-23 01:11:10
VSCodeVim/Vim
https://api.github.com/repos/VSCodeVim/Vim
closed
Investigate reducing startup activation time
area/performance
- Extension Name: vim - Extension Version: 0.15.7 - OS Version: Windows_NT x64 10.0.15063 - VSCode version: 1.25.1 We have written the needed data into your clipboard. Please paste: ![image](https://user-images.githubusercontent.com/6638644/43887188-9fc8efe0-9bf0-11e8-94d0-4932fe5fd658.png) why soooooooooooooooooooo slow???!!!!
True
Investigate reducing startup activation time - - Extension Name: vim - Extension Version: 0.15.7 - OS Version: Windows_NT x64 10.0.15063 - VSCode version: 1.25.1 We have written the needed data into your clipboard. Please paste: ![image](https://user-images.githubusercontent.com/6638644/43887188-9fc8efe0-9bf0-11e8-94d0-4932fe5fd658.png) why soooooooooooooooooooo slow???!!!!
non_process
investigate reducing startup activation time extension name vim extension version os version windows nt vscode version we have written the needed data into your clipboard please paste why soooooooooooooooooooo slow???!!!!
0
218,628
16,766,016,398
IssuesEvent
2021-06-14 08:54:06
jakobbossek/ecr3vis
https://api.github.com/repos/jakobbossek/ecr3vis
opened
Use mathjax for HTML output
documentation
We have many math formulas in the documentation. These are not rendered in the HTML RD files. Consider using [mathjaxr](http://cran.uni-muenster.de/web/packages/mathjaxr/mathjaxr.pdf)
1.0
Use mathjax for HTML output - We have many math formulas in the documentation. These are not rendered in the HTML RD files. Consider using [mathjaxr](http://cran.uni-muenster.de/web/packages/mathjaxr/mathjaxr.pdf)
non_process
use mathjax for html output we have many math formulas in the documentation these are not rendered in the html rd files consider using
0
349,799
31,831,824,270
IssuesEvent
2023-09-14 11:06:06
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachprod: implement "Reset" for aws and azure
C-enhancement X-stale A-roachprod no-issue-activity T-testeng O-cloudreport
Only GCE supports `Reset`--restarting a VM. As a result, some roachtests may be flaky (or incorrect) when executing outside of GCE. E.g., tpcc roachperf uses `Reset` [1] during each iteration of line search to determine an optimal number of warehouses. While the restart after each iteration is technically not required, it reduces noise. Other examples include failure-injection scenarios, e.g., restarting a VM at random. [1] https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/tests/tpcc.go#L1111 Epic: CRDB-10428 Jira issue: CRDB-13686
1.0
roachprod: implement "Reset" for aws and azure - Only GCE supports `Reset`--restarting a VM. As a result, some roachtests may be flaky (or incorrect) when executing outside of GCE. E.g., tpcc roachperf uses `Reset` [1] during each iteration of line search to determine an optimal number of warehouses. While the restart after each iteration is technically not required, it reduces noise. Other examples include failure-injection scenarios, e.g., restarting a VM at random. [1] https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/tests/tpcc.go#L1111 Epic: CRDB-10428 Jira issue: CRDB-13686
non_process
roachprod implement reset for aws and azure only gce supports reset restarting a vm as a result some roachtests may be flaky or incorrect when executing outside of gce e g tpcc roachperf uses reset during each iteration of line search to determine an optimal number of warehouses while the restart after each iteration is technically not required it reduces noise other examples include failure injection scenarios e g restarting a vm at random epic crdb jira issue crdb
0
19,448
25,727,174,600
IssuesEvent
2022-12-07 17:26:47
RobertCraigie/prisma-client-py
https://api.github.com/repos/RobertCraigie/prisma-client-py
opened
Could not connect to the Query Engine due to OSError [Errno 99]
bug/2-confirmed kind/bug process/candidate level/intermediate priority/high
<!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> A user has encountered this error: ``` OSError: [Errno 99] Cannot assign requested address File "httpcore/_exceptions.py", line 10, in map_exceptions yield File "httpcore/backends/sync.py", line 94, in connect_tcp sock = socket.create_connection( File "socket.py", line 844, in create_connection raise err File "socket.py", line 832, in create_connection sock.connect(sa) ``` ## How to reproduce <!-- Steps to reproduce the behavior: 1. Go to '...' 2. Change '....' 3. Run '....' 4. See error --> Not currently reproducible. Theorized cause is a race condition. ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> This should not crash.
1.0
Could not connect to the Query Engine due to OSError [Errno 99] - <!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> A user has encountered this error: ``` OSError: [Errno 99] Cannot assign requested address File "httpcore/_exceptions.py", line 10, in map_exceptions yield File "httpcore/backends/sync.py", line 94, in connect_tcp sock = socket.create_connection( File "socket.py", line 844, in create_connection raise err File "socket.py", line 832, in create_connection sock.connect(sa) ``` ## How to reproduce <!-- Steps to reproduce the behavior: 1. Go to '...' 2. Change '....' 3. Run '....' 4. See error --> Not currently reproducible. Theorized cause is a race condition. ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> This should not crash.
process
could not connect to the query engine due to oserror thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description a user has encountered this error oserror cannot assign requested address file httpcore exceptions py line in map exceptions yield file httpcore backends sync py line in connect tcp sock socket create connection file socket py line in create connection raise err file socket py line in create connection sock connect sa how to reproduce steps to reproduce the behavior go to change run see error not currently reproducible theorized cause is a race condition expected behavior this should not crash
1
834
3,297,247,142
IssuesEvent
2015-11-02 07:26:24
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.Diagnostics.Process missing environment variables.
System.Diagnostics.Process
/cc @Priya91, @pallavit, @joshfree, @stephentoub I am using Process to run a "Build.cmd", I have `echo %PATH%`, The result which run by C# is empty, at the same time, I tried add the system env vars again, it still not works. However, I double click the "Build.cmd", It can run the correct %PATH% ``` Process = new Process(); WorkingDirectory = FindDirectory(workingDirectory); var fileName = "cmd.exe"; if (OS.Current != OSType.Windows) { fileName = "bash"; } var arguments = "/c build.cmd"; if (OS.Current != OSType.Windows) { arguments = "./build.sh"; } Process.StartInfo = new ProcessStartInfo { FileName = fileName, Arguments = arguments, UseShellExecute = false, RedirectStandardError = true, RedirectStandardOutput = true, RedirectStandardInput = true, WorkingDirectory = WorkingDirectory }; ``` ... ``` var sysenv = Environment.GetEnvironmentVariables(); foreach(dynamic ev in sysenv) { #if DNXCORE50 if (Process.StartInfo.Environment[ev.Key] != null) Process.StartInfo.Environment[ev.Key] = Process.StartInfo.Environment[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value; else Process.StartInfo.Environment.Add(ev.Key, ev.Value); #else if (Process.StartInfo.EnvironmentVariables[ev.Key] != null) Process.StartInfo.EnvironmentVariables[ev.Key] = Process.StartInfo.EnvironmentVariables[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value; else Process.StartInfo.EnvironmentVariables.Add(ev.Key, ev.Value); #endif } ```
1.0
System.Diagnostics.Process missing environment variables. - /cc @Priya91, @pallavit, @joshfree, @stephentoub I am using Process to run a "Build.cmd", I have `echo %PATH%`, The result which run by C# is empty, at the same time, I tried add the system env vars again, it still not works. However, I double click the "Build.cmd", It can run the correct %PATH% ``` Process = new Process(); WorkingDirectory = FindDirectory(workingDirectory); var fileName = "cmd.exe"; if (OS.Current != OSType.Windows) { fileName = "bash"; } var arguments = "/c build.cmd"; if (OS.Current != OSType.Windows) { arguments = "./build.sh"; } Process.StartInfo = new ProcessStartInfo { FileName = fileName, Arguments = arguments, UseShellExecute = false, RedirectStandardError = true, RedirectStandardOutput = true, RedirectStandardInput = true, WorkingDirectory = WorkingDirectory }; ``` ... ``` var sysenv = Environment.GetEnvironmentVariables(); foreach(dynamic ev in sysenv) { #if DNXCORE50 if (Process.StartInfo.Environment[ev.Key] != null) Process.StartInfo.Environment[ev.Key] = Process.StartInfo.Environment[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value; else Process.StartInfo.Environment.Add(ev.Key, ev.Value); #else if (Process.StartInfo.EnvironmentVariables[ev.Key] != null) Process.StartInfo.EnvironmentVariables[ev.Key] = Process.StartInfo.EnvironmentVariables[ev.Key].TrimEnd(' ').TrimEnd(';') + ";" + ev.Value; else Process.StartInfo.EnvironmentVariables.Add(ev.Key, ev.Value); #endif } ```
process
system diagnostics process missing environment variables cc pallavit joshfree stephentoub i am using process to run a build cmd i have echo path the result which run by c is empty at the same time i tried add the system env vars again it still not works however i double click the build cmd it can run the correct path process new process workingdirectory finddirectory workingdirectory var filename cmd exe if os current ostype windows filename bash var arguments c build cmd if os current ostype windows arguments build sh process startinfo new processstartinfo filename filename arguments arguments useshellexecute false redirectstandarderror true redirectstandardoutput true redirectstandardinput true workingdirectory workingdirectory var sysenv environment getenvironmentvariables foreach dynamic ev in sysenv if if process startinfo environment null process startinfo environment process startinfo environment trimend trimend ev value else process startinfo environment add ev key ev value else if process startinfo environmentvariables null process startinfo environmentvariables process startinfo environmentvariables trimend trimend ev value else process startinfo environmentvariables add ev key ev value endif
1
14,234
17,154,611,056
IssuesEvent
2021-07-14 04:15:51
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
opened
Provision for app upgrades
Android P1 Participant datastore Process: Enhancement iOS
1. Notify app users if a new version of the mobile app is available on the app stores. The notification should redirect to the app store page. 2. Have a provision to configure forced vs. optional upgrade app behavior. If there is a new app version available, users see a message when they launch or visit the app, asking them to upgrade the app before continuing to use it. This can be a mandatory or optional step depending on a server-side configuration. 3. Users should be able to resume app usage smoothly post app upgrade. Requiring them to sign in again is acceptable if there are server-side changes that necessitate this.
1.0
Provision for app upgrades - 1. Notify app users if a new version of the mobile app is available on the app stores. The notification should redirect to the app store page. 2. Have a provision to configure forced vs. optional upgrade app behavior. If there is a new app version available, users see a message when they launch or visit the app, asking them to upgrade the app before continuing to use it. This can be a mandatory or optional step depending on a server-side configuration. 3. Users should be able to resume app usage smoothly post app upgrade. Requiring them to sign in again is acceptable if there are server-side changes that necessitate this.
process
provision for app upgrades notify app users if a new version of the mobile app is available on the app stores the notification should redirect to the app store page have a provision to configure forced vs optional upgrade app behavior if there is a new app version available users see a message when they launch or visit the app asking them to upgrade the app before continuing to use it this can be a mandatory or optional step depending on a server side configuration users should be able to resume app usage smoothly post app upgrade requiring them to sign in again is acceptable if there are server side changes that necessitate this
1
502,276
14,543,479,091
IssuesEvent
2020-12-15 16:55:44
zulip/zulip-mobile
https://api.github.com/repos/zulip/zulip-mobile
closed
Add mobile support for new `user_avatar_url_field_optional` client capability
P1 high-priority a-avatar api migrations
To resolve https://github.com/zulip/zulip/pull/15287, we're introducing a new client_capability that should allow the mobile app to have good performance when talking to servers with thousands of long-term-idle users (and email_address_visibility configured; that last bit being relevant mainly in that our previous attempt at solving this problem. `client_gravatar` feature works only with EMAIL_ADDRESS_VISIBILITY_EVERYONE), since the client needs real email addresses to compute gravatar hashes. One should be able to prototype today against https://github.com/zulip/zulip/pull/15359; it should work aside from having the wrong capability name. I expect that to get cleaned up and this to merge in the next few days. Tagging as a priority since this issue makes chat.zulip.org a lot slower to load on mobile.
1.0
Add mobile support for new `user_avatar_url_field_optional` client capability - To resolve https://github.com/zulip/zulip/pull/15287, we're introducing a new client_capability that should allow the mobile app to have good performance when talking to servers with thousands of long-term-idle users (and email_address_visibility configured; that last bit being relevant mainly in that our previous attempt at solving this problem. `client_gravatar` feature works only with EMAIL_ADDRESS_VISIBILITY_EVERYONE), since the client needs real email addresses to compute gravatar hashes. One should be able to prototype today against https://github.com/zulip/zulip/pull/15359; it should work aside from having the wrong capability name. I expect that to get cleaned up and this to merge in the next few days. Tagging as a priority since this issue makes chat.zulip.org a lot slower to load on mobile.
non_process
add mobile support for new user avatar url field optional client capability to resolve we re introducing a new client capability that should allow the mobile app to have good performance when talking to servers with thousands of long term idle users and email address visibility configured that last bit being relevant mainly in that our previous attempt at solving this problem client gravatar feature works only with email address visibility everyone since the client needs real email addresses to compute gravatar hashes one should be able to prototype today against it should work aside from having the wrong capability name i expect that to get cleaned up and this to merge in the next few days tagging as a priority since this issue makes chat zulip org a lot slower to load on mobile
0
71,943
9,545,095,416
IssuesEvent
2019-05-01 16:02:33
CosmiQ/cw-nets
https://api.github.com/repos/CosmiQ/cw-nets
closed
Augmentation docs
Difficulty: Easy Priority: High Status: On Hold Type: Documentation
After completing augmentation implementation (#35) we need to document it. Documentation components: [ ] list of available augmentations [ ] set of augmentations only compatible with 3-channel imagery [ ] set of required arguments for each augmentation [ ] instructions and examples for using the `cw_nets.data.transform` submodule, including yaml config formatting
1.0
Augmentation docs - After completing augmentation implementation (#35) we need to document it. Documentation components: [ ] list of available augmentations [ ] set of augmentations only compatible with 3-channel imagery [ ] set of required arguments for each augmentation [ ] instructions and examples for using the `cw_nets.data.transform` submodule, including yaml config formatting
non_process
augmentation docs after completing augmentation implementation we need to document it documentation components list of available augmentations set of augmentations only compatible with channel imagery set of required arguments for each augmentation instructions and examples for using the cw nets data transform submodule including yaml config formatting
0
7,359
10,509,141,261
IssuesEvent
2019-09-27 10:14:32
prisma/studio
https://api.github.com/repos/prisma/studio
opened
Reload unintuitive
kind/improvement process/candidate
## Reload in Table Plus ![image](https://user-images.githubusercontent.com/1094804/65761905-16d87b80-e120-11e9-8bf8-2c9f663d4c70.png) ## Reload in Studio ![image](https://user-images.githubusercontent.com/1094804/65761940-28218800-e120-11e9-8365-11ae24967aba.png) As a user, who just wants the data to reload, without knowing that Photon is used under the hood, I don't know, what `Run` means in this context. I just want the UI to reload. I suggest either calling it `Reload` or replacing it with a reload icon.
1.0
Reload unintuitive - ## Reload in Table Plus ![image](https://user-images.githubusercontent.com/1094804/65761905-16d87b80-e120-11e9-8bf8-2c9f663d4c70.png) ## Reload in Studio ![image](https://user-images.githubusercontent.com/1094804/65761940-28218800-e120-11e9-8365-11ae24967aba.png) As a user, who just wants the data to reload, without knowing that Photon is used under the hood, I don't know, what `Run` means in this context. I just want the UI to reload. I suggest either calling it `Reload` or replacing it with a reload icon.
process
reload unintuitive reload in table plus reload in studio as a user who just wants the data to reload without knowing that photon is used under the hood i don t know what run means in this context i just want the ui to reload i suggest either calling it reload or replacing it with a reload icon
1
3,233
6,289,280,115
IssuesEvent
2017-07-19 18:51:21
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess fails with NRE on UAP
area-System.Diagnostics.Process bug os-windows-uwp
(Test case will be added soon, creating issue so that I can disable that in the PR) ``` ERROR: System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess [FAIL] System.AggregateException : One or more errors occurred. (Object reference not set to an instance of an object.) (Object reference not s et to an instance of an object.) ---- System.NullReferenceException : Object reference not set to an instance of an object. ---- System.NullReferenceException : Object reference not set to an instance of an object. Stack Trace: ----- Inner Stack Trace #1 (System.NullReferenceException) ----- at System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess() ----- Inner Stack Trace #2 (System.NullReferenceException) ----- at System.Diagnostics.Tests.ProcessTestBase.Dispose(Boolean disposing) at System.IO.FileCleanupTestBase.Dispose() at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, Can cellationTokenSource cancellationTokenSource) ```
1.0
System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess fails with NRE on UAP - (Test case will be added soon, creating issue so that I can disable that in the PR) ``` ERROR: System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess [FAIL] System.AggregateException : One or more errors occurred. (Object reference not set to an instance of an object.) (Object reference not s et to an instance of an object.) ---- System.NullReferenceException : Object reference not set to an instance of an object. ---- System.NullReferenceException : Object reference not set to an instance of an object. Stack Trace: ----- Inner Stack Trace #1 (System.NullReferenceException) ----- at System.Diagnostics.Tests.ProcessWaitingTests.WaitForPeerProcess() ----- Inner Stack Trace #2 (System.NullReferenceException) ----- at System.Diagnostics.Tests.ProcessTestBase.Dispose(Boolean disposing) at System.IO.FileCleanupTestBase.Dispose() at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, Can cellationTokenSource cancellationTokenSource) ```
process
system diagnostics tests processwaitingtests waitforpeerprocess fails with nre on uap test case will be added soon creating issue so that i can disable that in the pr error system diagnostics tests processwaitingtests waitforpeerprocess system aggregateexception one or more errors occurred object reference not set to an instance of an object object reference not s et to an instance of an object system nullreferenceexception object reference not set to an instance of an object system nullreferenceexception object reference not set to an instance of an object stack trace inner stack trace system nullreferenceexception at system diagnostics tests processwaitingtests waitforpeerprocess inner stack trace system nullreferenceexception at system diagnostics tests processtestbase dispose boolean disposing at system io filecleanuptestbase dispose at reflectionabstractionextensions disposetestclass itest test object testclass imessagebus messagebus executiontimer timer can cellationtokensource cancellationtokensource
1
125,894
17,861,285,972
IssuesEvent
2021-09-06 01:06:02
bsbtd/Teste
https://api.github.com/repos/bsbtd/Teste
opened
CVE-2021-23437 (High) detected in Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl
security vulnerability
## CVE-2021-23437 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: Teste/pytorch-metric-learning</p> <p>Path to vulnerable library: Teste/pytorch-metric-learning</p> <p> Dependency Hierarchy: - :x: **Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package pillow from 0 and before 8.3.2 are vulnerable to Regular Expression Denial of Service (ReDoS) via the getrgb function. <p>Publish Date: 2021-09-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23437>CVE-2021-23437</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html">https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html</a></p> <p>Release Date: 2021-09-03</p> <p>Fix Resolution: Pillow - 8.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23437 (High) detected in Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl - ## CVE-2021-23437 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/e0/50/8e78e6f62ffa50d6ca95c281d5a2819bef66d023ac1b723e253de5bda9c5/Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: Teste/pytorch-metric-learning</p> <p>Path to vulnerable library: Teste/pytorch-metric-learning</p> <p> Dependency Hierarchy: - :x: **Pillow-7.1.2-cp36-cp36m-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package pillow from 0 and before 8.3.2 are vulnerable to Regular Expression Denial of Service (ReDoS) via the getrgb function. <p>Publish Date: 2021-09-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23437>CVE-2021-23437</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html">https://pillow.readthedocs.io/en/stable/releasenotes/8.3.2.html</a></p> <p>Release Date: 2021-09-03</p> <p>Fix Resolution: Pillow - 8.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file teste pytorch metric learning path to vulnerable library teste pytorch metric learning dependency hierarchy x pillow whl vulnerable library vulnerability details the package pillow from and before are vulnerable to regular expression denial of service redos via the getrgb function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pillow step up your open source security game with whitesource
0
22,537
31,707,807,504
IssuesEvent
2023-09-09 00:18:23
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Can’t scale up a node pool because of a failing scheduling predicate
bug process
### Description Seen a couple times where GKE node pools can't scale because of a failing scheduling predicate. Logs indicate it was promtail and node exporter daemonset. These daemonset pods could not get scheduled because other items on the node had higher priority. This would be fine by itself since these pods are not the highest priority, but if it causes the node to not scale then it needs to be addressed. ### Steps to reproduce Monitor GKE with real life workloads. ### Additional context _No response_ ### Hedera network other ### Version 0.87.0 ### Operating system None
1.0
Can’t scale up a node pool because of a failing scheduling predicate - ### Description Seen a couple times where GKE node pools can't scale because of a failing scheduling predicate. Logs indicate it was promtail and node exporter daemonset. These daemonset pods could not get scheduled because other items on the node had higher priority. This would be fine by itself since these pods are not the highest priority, but if it causes the node to not scale then it needs to be addressed. ### Steps to reproduce Monitor GKE with real life workloads. ### Additional context _No response_ ### Hedera network other ### Version 0.87.0 ### Operating system None
process
can’t scale up a node pool because of a failing scheduling predicate description seen a couple times where gke node pools can t scale because of a failing scheduling predicate logs indicate it was promtail and node exporter daemonset these daemonset pods could not get scheduled because other items on the node had higher priority this would be fine by itself since these pods are not the highest priority but if it causes the node to not scale then it needs to be addressed steps to reproduce monitor gke with real life workloads additional context no response hedera network other version operating system none
1
212,093
16,472,784,182
IssuesEvent
2021-05-23 18:53:41
truecharts/apps
https://api.github.com/repos/truecharts/apps
closed
Adapt for persitance.emptyDir to persistance.emptyDir.enabled
documentation enhancement good first issue
**Is your feature request related to a problem? Please describe.** common 4.0.0 gave persistance.emptyDir it's own sub parameter `enabled` **Describe the solution you'd like** Adapt current docs and charts accordingly.
1.0
Adapt for persitance.emptyDir to persistance.emptyDir.enabled - **Is your feature request related to a problem? Please describe.** common 4.0.0 gave persistance.emptyDir it's own sub parameter `enabled` **Describe the solution you'd like** Adapt current docs and charts accordingly.
non_process
adapt for persitance emptydir to persistance emptydir enabled is your feature request related to a problem please describe common gave persistance emptydir it s own sub parameter enabled describe the solution you d like adapt current docs and charts accordingly
0
3,714
6,732,600,623
IssuesEvent
2017-10-18 12:10:11
lockedata/rcms
https://api.github.com/repos/lockedata/rcms
opened
Manage attendees
conference team osem processes
## Detailed task - Monitor sales - Modify a registration e.g. issue a refund - Send an email to attendees ## Assessing the task Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks. Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback. ## Extra Info - Site: [osem](https://intense-shore-93790.herokuapp.com/) - System documentation: [osem docs](http://osem.io/) - Role: Conference team - Area: Processes
1.0
Manage attendees - ## Detailed task - Monitor sales - Modify a registration e.g. issue a refund - Send an email to attendees ## Assessing the task Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks. Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback. ## Extra Info - Site: [osem](https://intense-shore-93790.herokuapp.com/) - System documentation: [osem docs](http://osem.io/) - Role: Conference team - Area: Processes
process
manage attendees detailed task monitor sales modify a registration e g issue a refund send an email to attendees assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role conference team area processes
1
18,530
24,552,656,670
IssuesEvent
2022-10-12 13:43:18
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] My Account > Toggle buttons should not be functional in offline behaviour
Bug P1 iOS Process: Fixed Process: Tested dev
**Steps:** 1. Install the app 2. Sign in/signup 3. Navigated to My account 4. Switch off the internet 6. Now enable/disable toggle buttons like 'Receive push notifications?' or 'Receive study activity reminders?' 7. Switch on the internet 8. Navigate to different screen and come back to my account 9. Observe buttons enabled/disabled fails to retain **Actual:** Button functionality fails to retain once navigated from different screen to my account **Expected:** Toggle buttons should not be functional in offline behaviour Refer video https://user-images.githubusercontent.com/60386291/179733053-e6086ff4-ba51-42e4-86a2-6d3f3802a6e2.MOV
2.0
[iOS] [Offline indicator] My Account > Toggle buttons should not be functional in offline behaviour - **Steps:** 1. Install the app 2. Sign in/signup 3. Navigated to My account 4. Switch off the internet 6. Now enable/disable toggle buttons like 'Receive push notifications?' or 'Receive study activity reminders?' 7. Switch on the internet 8. Navigate to different screen and come back to my account 9. Observe buttons enabled/disabled fails to retain **Actual:** Button functionality fails to retain once navigated from different screen to my account **Expected:** Toggle buttons should not be functional in offline behaviour Refer video https://user-images.githubusercontent.com/60386291/179733053-e6086ff4-ba51-42e4-86a2-6d3f3802a6e2.MOV
process
my account toggle buttons should not be functional in offline behaviour steps install the app sign in signup navigated to my account switch off the internet now enable disable toggle buttons like receive push notifications or receive study activity reminders switch on the internet navigate to different screen and come back to my account observe buttons enabled disabled fails to retain actual button functionality fails to retain once navigated from different screen to my account expected toggle buttons should not be functional in offline behaviour refer video
1
70,963
13,564,458,780
IssuesEvent
2020-09-18 10:04:28
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
Death desync after moving sub away from spawnpoint
Bug Code High prio Networking
- [X] I have searched the issue tracker to check if the issue has already been reported. **Description** ~After respawning in a shuttle on a mission~ After moving away from spawnpoint there will be massive death-related desync. When you die server-side you will be locked in place instead. When you are knocked out you can't `give up` and spectate. **Steps To Reproduce** (Original, see below for a non-shuttle related one, though this one works too) 1. Host server (dedicated or p2p). 2. Start mission. 3. Kill yourself (for example `kill` command). 4. Respawn in a shuttle (`respawnnow` command). 5. Kill yourself again (`kill` command). 6. Notice that you will die on the server, but you will still appear very alive for some time but can't move or give up. **Version** 0.10.505.0 - 0.10.5 **Additional information** I'm pretty sure it did not happen in the previous minor unstable version. EDIT: Update description, version and steps to reproduce with newly found information.
1.0
Death desync after moving sub away from spawnpoint - - [X] I have searched the issue tracker to check if the issue has already been reported. **Description** ~After respawning in a shuttle on a mission~ After moving away from spawnpoint there will be massive death-related desync. When you die server-side you will be locked in place instead. When you are knocked out you can't `give up` and spectate. **Steps To Reproduce** (Original, see below for a non-shuttle related one, though this one works too) 1. Host server (dedicated or p2p). 2. Start mission. 3. Kill yourself (for example `kill` command). 4. Respawn in a shuttle (`respawnnow` command). 5. Kill yourself again (`kill` command). 6. Notice that you will die on the server, but you will still appear very alive for some time but can't move or give up. **Version** 0.10.505.0 - 0.10.5 **Additional information** I'm pretty sure it did not happen in the previous minor unstable version. EDIT: Update description, version and steps to reproduce with newly found information.
non_process
death desync after moving sub away from spawnpoint i have searched the issue tracker to check if the issue has already been reported description after respawning in a shuttle on a mission after moving away from spawnpoint there will be massive death related desync when you die server side you will be locked in place instead when you are knocked out you can t give up and spectate steps to reproduce original see below for a non shuttle related one though this one works too host server dedicated or start mission kill yourself for example kill command respawn in a shuttle respawnnow command kill yourself again kill command notice that you will die on the server but you will still appear very alive for some time but can t move or give up version additional information i m pretty sure it did not happen in the previous minor unstable version edit update description version and steps to reproduce with newly found information
0
17,944
5,535,467,504
IssuesEvent
2017-03-21 17:25:46
phetsims/masses-and-springs
https://api.github.com/repos/phetsims/masses-and-springs
opened
Factor out duplicated code in line creation
dev:code-review
During #36 I saw this code in IndicatorVisibilityControlPanel: ```js // Lines added for reference in panel var greenLine = new Line( 0, 0, LINE_LENGTH, 0, { stroke: 'rgb(93, 191, 142)', lineDash: [ 6, 2.5 ], lineWidth: 2.0, cursor: 'pointer', tandem: tandem.createTandem( 'greenLine' ) } ); var blueLine = new Line( 0, 0, LINE_LENGTH, 0, { stroke: 'rgb(65,66,232)', lineDash: [ 6, 2.5 ], lineWidth: 2.0, cursor: 'pointer', tandem: tandem.createTandem( 'blueLine' ) } ); var redLine = new Line( 0, 0, LINE_LENGTH, 0, { stroke: 'red', lineDash: [ 6, 2.5 ], lineWidth: 2.0, cursor: 'pointer', tandem: tandem.createTandem( 'redLine' ) } ); ``` I recommend factoring out a function so the lines can be created like this: ```js var greenLine = createLine('rgb(93, 191, 142)',tandem.createTandem('greenLine')); ```
1.0
Factor out duplicated code in line creation - During #36 I saw this code in IndicatorVisibilityControlPanel: ```js // Lines added for reference in panel var greenLine = new Line( 0, 0, LINE_LENGTH, 0, { stroke: 'rgb(93, 191, 142)', lineDash: [ 6, 2.5 ], lineWidth: 2.0, cursor: 'pointer', tandem: tandem.createTandem( 'greenLine' ) } ); var blueLine = new Line( 0, 0, LINE_LENGTH, 0, { stroke: 'rgb(65,66,232)', lineDash: [ 6, 2.5 ], lineWidth: 2.0, cursor: 'pointer', tandem: tandem.createTandem( 'blueLine' ) } ); var redLine = new Line( 0, 0, LINE_LENGTH, 0, { stroke: 'red', lineDash: [ 6, 2.5 ], lineWidth: 2.0, cursor: 'pointer', tandem: tandem.createTandem( 'redLine' ) } ); ``` I recommend factoring out a function so the lines can be created like this: ```js var greenLine = createLine('rgb(93, 191, 142)',tandem.createTandem('greenLine')); ```
non_process
factor out duplicated code in line creation during i saw this code in indicatorvisibilitycontrolpanel js lines added for reference in panel var greenline new line line length stroke rgb linedash linewidth cursor pointer tandem tandem createtandem greenline var blueline new line line length stroke rgb linedash linewidth cursor pointer tandem tandem createtandem blueline var redline new line line length stroke red linedash linewidth cursor pointer tandem tandem createtandem redline i recommend factoring out a function so the lines can be created like this js var greenline createline rgb tandem createtandem greenline
0
5,382
8,211,044,554
IssuesEvent
2018-09-04 12:45:58
linnovate/root
https://api.github.com/repos/linnovate/root
closed
Document: filtering by favorite not working
Process bug
@abrahamos open a few documents. set one of them as a favorite. click on filtering by favorite. all the documents still there. ![temp](https://user-images.githubusercontent.com/31100069/44717880-fece9c80-aac6-11e8-9911-e1002cf0cb29.PNG)
1.0
Document: filtering by favorite not working - @abrahamos open a few documents. set one of them as a favorite. click on filtering by favorite. all the documents still there. ![temp](https://user-images.githubusercontent.com/31100069/44717880-fece9c80-aac6-11e8-9911-e1002cf0cb29.PNG)
process
document filtering by favorite not working abrahamos open a few documents set one of them as a favorite click on filtering by favorite all the documents still there
1
2,447
5,226,087,836
IssuesEvent
2017-01-27 20:14:34
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
Run WebGL tests in CI
dev process enhancement priority
As discussed with @mramato offline: - Replace all read pixels expectations with a function that can have a no-op expectation when the tests are ran with a "no WebGL" flag, e.g., ``` javascript expect(scene.renderForSpecs()).toEqual([0, 0, 0, 255]); ``` becomes ``` javascript scene.expectRenderForSpecs([0, 0, 0, 255]); ``` - Replace the object returned by `getContext` with a mock object with GL functions that are no-ops, `function() {}`, except for `getExtension`, which should return mocked objects for the extensions we care about. - Likewise, all `gl.get*` functions should be mocked to return reasonable values. - Run the tests and fix things I forgot. This should only take a few hours and will be more reliable than [mesa](https://github.com/AnalyticalGraphicsInc/cesium/compare/mesa).
1.0
Run WebGL tests in CI - As discussed with @mramato offline: - Replace all read pixels expectations with a function that can have a no-op expectation when the tests are ran with a "no WebGL" flag, e.g., ``` javascript expect(scene.renderForSpecs()).toEqual([0, 0, 0, 255]); ``` becomes ``` javascript scene.expectRenderForSpecs([0, 0, 0, 255]); ``` - Replace the object returned by `getContext` with a mock object with GL functions that are no-ops, `function() {}`, except for `getExtension`, which should return mocked objects for the extensions we care about. - Likewise, all `gl.get*` functions should be mocked to return reasonable values. - Run the tests and fix things I forgot. This should only take a few hours and will be more reliable than [mesa](https://github.com/AnalyticalGraphicsInc/cesium/compare/mesa).
process
run webgl tests in ci as discussed with mramato offline replace all read pixels expectations with a function that can have a no op expectation when the tests are ran with a no webgl flag e g javascript expect scene renderforspecs toequal becomes javascript scene expectrenderforspecs replace the object returned by getcontext with a mock object with gl functions that are no ops function except for getextension which should return mocked objects for the extensions we care about likewise all gl get functions should be mocked to return reasonable values run the tests and fix things i forgot this should only take a few hours and will be more reliable than
1
136,522
11,049,379,521
IssuesEvent
2019-12-09 23:32:05
MangopearUK/European-Boating-Association--Theme
https://api.github.com/repos/MangopearUK/European-Boating-Association--Theme
closed
Test & audit: EBA subscription rate increase for budget year 2020
Testing: second round
Page URL: https://eba.eu.com/membership/secretariat-announcements/eba-subscription-rate-increase-for-budget-year-2020/ ## Table of contents - [x] **Task 1:** Perform automated audits _(10 tasks)_ - [x] **Task 2:** Manual standards & accessibility tests _(61 tasks)_ - [x] **Task 3:** Breakpoint testing _(15 tasks)_ - [x] **Task 4:** Re-run automated audits _(10 tasks)_ ## 1: Perform automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors ## 2: Manual standards & accessibility tests _(61 tasks)_ ### Forms - [x] Give all form elements permanently visible labels - [x] Place labels above form elements - [x] Mark invalid fields clearly and provide associated error messages - [x] Make forms as short as possible; offer shortcuts like autocompleting the address using the postcode - [x] Ensure all form fields have the correct requried state - [x] Provide status and error messages as WAI-ARIA live regions ### Readability of content - [x] Ensure page has good grammar - [x] Ensure page content has been spell-checked - [x] Make sure headings are in logical order - [x] Ensure the same content is available across different devices and platforms - [x] Begin long, multi-section documents with a table of contents ### Presentation - [x] Make sure all content is formatted correctly - [x] Avoid all-caps text - [x] Make sure data tables wider than their container can be scrolled horizontally - [x] Use the same design patterns to solve the same problems - [x] Do not mark up subheadings/straplines with separate heading elements ### Links & buttons #### Links - [x] Check all links to ensure they work - [x] Check all links to third party websites use `rel="noopener"` - [x] Make sure the purpose of a link is clearly described: "read more" vs. "read more about accessibility" - [x] Provide a skip link if necessary - [x] Underline links — at least in body copy - [x] Warn users of links that have unusual behaviors, like linking off-site, or loading a new tab (i.e. aria-label) #### Buttons - [x] Ensure primary calls to action are easy to recognize and reach - [x] Provide clear, unambiguous focus styles - [x] Ensure states (pressed, expanded, invalid, etc) are communicated to assistive software - [x] Ensure disabled controls are not focusable - [x] Make sure controls within hidden content are not focusable - [x] Provide large touch "targets" for interactive elements - [x] Make controls look like controls; give them strong perceived affordance - [x] Use well-established, therefore recognizable, icons and symbols ### Assistive technology - [x] Ensure content is not obscured through zooming - [x] Support Windows high contrast mode (use images, not background images) - [x] Provide alternative text for salient images - [x] Make scrollable elements focusable for keyboard users - [x] Ensure keyboard focus order is logical regarding visual layout - [x] Match semantics to behavior for assistive technology users - [x] Provide a default language and use lang="[ISO code]" for subsections in different languages - [x] Inform the user when there are important changes to the application state - [x] Do not hijack standard scrolling behavior - [x] Do not instate "infinite scroll" by default; provide buttons to load more items ### General accessibility - [x] Make sure text and background colors contrast sufficiently - [x] Do not rely on color for differentiation of visual elements - [x] Avoid images of text — text that cannot be translated, selected, or understood by assistive tech - [x] Provide a print stylesheet - [x] Honour requests to remove animation via the prefers-reduced-motion media query ### SEO - [x] Ensure all pages have appropriate title - [x] Ensure all pages have meta descriptions - [x] Make content easier to find and improve search results with structured data [Read more](https://developers.google.com/search/docs/guides/prototype) - [x] Check whether page should be appearing in sitemap - [x] Make sure page has Facebook and Twitter large image previews set correctly - [x] Check canonical links for page - [x] Mark as cornerstone content? ### Performance - [x] Ensure all CSS assets are minified and concatenated - [x] Ensure all JS assets are minified and concatenated - [x] Ensure all images are compressed - [x] Where possible, remove redundant code - [x] Ensure all SVG assets have been optimised - [x] Make sure styles and scripts are not render blocking - [x] Ensure large image assets are lazy loaded ### Other - [x] Make sure all content belongs to a landmark element - [x] Provide a manifest.json file for identifiable homescreen entries ## 3: Breakpoint testing _(15 tasks)_ ### Desktop - [x] Provide a full screenshot of **1920px** wide page - [x] Provide a full screenshot of **1500px** wide page - [x] Provide a full screenshot of **1280px** wide page - [x] Provide a full screenshot of **1024px** wide page ### Tablet - [x] Provide a full screenshot of **960px** wide page - [x] Provide a full screenshot of **800px** wide page - [x] Provide a full screenshot of **760px** wide page - [x] Provide a full screenshot of **650px** wide page ### Mobile - [x] Provide a full screenshot of **600px** wide page - [x] Provide a full screenshot of **500px** wide page - [x] Provide a full screenshot of **450px** wide page - [x] Provide a full screenshot of **380px** wide page - [x] Provide a full screenshot of **320px** wide page - [x] Provide a full screenshot of **280px** wide page - [x] Provide a full screenshot of **250px** wide page ## 4: Re-run automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors
1.0
Test & audit: EBA subscription rate increase for budget year 2020 - Page URL: https://eba.eu.com/membership/secretariat-announcements/eba-subscription-rate-increase-for-budget-year-2020/ ## Table of contents - [x] **Task 1:** Perform automated audits _(10 tasks)_ - [x] **Task 2:** Manual standards & accessibility tests _(61 tasks)_ - [x] **Task 3:** Breakpoint testing _(15 tasks)_ - [x] **Task 4:** Re-run automated audits _(10 tasks)_ ## 1: Perform automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors ## 2: Manual standards & accessibility tests _(61 tasks)_ ### Forms - [x] Give all form elements permanently visible labels - [x] Place labels above form elements - [x] Mark invalid fields clearly and provide associated error messages - [x] Make forms as short as possible; offer shortcuts like autocompleting the address using the postcode - [x] Ensure all form fields have the correct requried state - [x] Provide status and error messages as WAI-ARIA live regions ### Readability of content - [x] Ensure page has good grammar - [x] Ensure page content has been spell-checked - [x] Make sure headings are in logical order - [x] Ensure the same content is available across different devices and platforms - [x] Begin long, multi-section documents with a table of contents ### Presentation - [x] Make sure all content is formatted correctly - [x] Avoid all-caps text - [x] Make sure data tables wider than their container can be scrolled horizontally - [x] Use the same design patterns to solve the same problems - [x] Do not mark up subheadings/straplines with separate heading elements ### Links & buttons #### Links - [x] Check all links to ensure they work - [x] Check all links to third party websites use `rel="noopener"` - [x] Make sure the purpose of a link is clearly described: "read more" vs. "read more about accessibility" - [x] Provide a skip link if necessary - [x] Underline links — at least in body copy - [x] Warn users of links that have unusual behaviors, like linking off-site, or loading a new tab (i.e. aria-label) #### Buttons - [x] Ensure primary calls to action are easy to recognize and reach - [x] Provide clear, unambiguous focus styles - [x] Ensure states (pressed, expanded, invalid, etc) are communicated to assistive software - [x] Ensure disabled controls are not focusable - [x] Make sure controls within hidden content are not focusable - [x] Provide large touch "targets" for interactive elements - [x] Make controls look like controls; give them strong perceived affordance - [x] Use well-established, therefore recognizable, icons and symbols ### Assistive technology - [x] Ensure content is not obscured through zooming - [x] Support Windows high contrast mode (use images, not background images) - [x] Provide alternative text for salient images - [x] Make scrollable elements focusable for keyboard users - [x] Ensure keyboard focus order is logical regarding visual layout - [x] Match semantics to behavior for assistive technology users - [x] Provide a default language and use lang="[ISO code]" for subsections in different languages - [x] Inform the user when there are important changes to the application state - [x] Do not hijack standard scrolling behavior - [x] Do not instate "infinite scroll" by default; provide buttons to load more items ### General accessibility - [x] Make sure text and background colors contrast sufficiently - [x] Do not rely on color for differentiation of visual elements - [x] Avoid images of text — text that cannot be translated, selected, or understood by assistive tech - [x] Provide a print stylesheet - [x] Honour requests to remove animation via the prefers-reduced-motion media query ### SEO - [x] Ensure all pages have appropriate title - [x] Ensure all pages have meta descriptions - [x] Make content easier to find and improve search results with structured data [Read more](https://developers.google.com/search/docs/guides/prototype) - [x] Check whether page should be appearing in sitemap - [x] Make sure page has Facebook and Twitter large image previews set correctly - [x] Check canonical links for page - [x] Mark as cornerstone content? ### Performance - [x] Ensure all CSS assets are minified and concatenated - [x] Ensure all JS assets are minified and concatenated - [x] Ensure all images are compressed - [x] Where possible, remove redundant code - [x] Ensure all SVG assets have been optimised - [x] Make sure styles and scripts are not render blocking - [x] Ensure large image assets are lazy loaded ### Other - [x] Make sure all content belongs to a landmark element - [x] Provide a manifest.json file for identifiable homescreen entries ## 3: Breakpoint testing _(15 tasks)_ ### Desktop - [x] Provide a full screenshot of **1920px** wide page - [x] Provide a full screenshot of **1500px** wide page - [x] Provide a full screenshot of **1280px** wide page - [x] Provide a full screenshot of **1024px** wide page ### Tablet - [x] Provide a full screenshot of **960px** wide page - [x] Provide a full screenshot of **800px** wide page - [x] Provide a full screenshot of **760px** wide page - [x] Provide a full screenshot of **650px** wide page ### Mobile - [x] Provide a full screenshot of **600px** wide page - [x] Provide a full screenshot of **500px** wide page - [x] Provide a full screenshot of **450px** wide page - [x] Provide a full screenshot of **380px** wide page - [x] Provide a full screenshot of **320px** wide page - [x] Provide a full screenshot of **280px** wide page - [x] Provide a full screenshot of **250px** wide page ## 4: Re-run automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors
non_process
test audit eba subscription rate increase for budget year page url table of contents task perform automated audits tasks task manual standards accessibility tests tasks task breakpoint testing tasks task re run automated audits tasks perform automated audits tasks lighthouse run accessibility audit in lighthouse using incognito tab run performance audit in lighthouse using incognito tab run best practices audit in lighthouse using incognito tab run seo audit in lighthouse using incognito tab run pwa audit in lighthouse using incognito tab pingdom run full audit of the the page s performance in pingdom browser s console check chrome s console for errors log results of audits screenshot snapshot of the lighthouse audits upload pdf of detailed lighthouse reports provide a screenshot of any console errors manual standards accessibility tests tasks forms give all form elements permanently visible labels place labels above form elements mark invalid fields clearly and provide associated error messages make forms as short as possible offer shortcuts like autocompleting the address using the postcode ensure all form fields have the correct requried state provide status and error messages as wai aria live regions readability of content ensure page has good grammar ensure page content has been spell checked make sure headings are in logical order ensure the same content is available across different devices and platforms begin long multi section documents with a table of contents presentation make sure all content is formatted correctly avoid all caps text make sure data tables wider than their container can be scrolled horizontally use the same design patterns to solve the same problems do not mark up subheadings straplines with separate heading elements links buttons links check all links to ensure they work check all links to third party websites use rel noopener make sure the purpose of a link is clearly described read more vs read more about accessibility provide a skip link if necessary underline links — at least in body copy warn users of links that have unusual behaviors like linking off site or loading a new tab i e aria label buttons ensure primary calls to action are easy to recognize and reach provide clear unambiguous focus styles ensure states pressed expanded invalid etc are communicated to assistive software ensure disabled controls are not focusable make sure controls within hidden content are not focusable provide large touch targets for interactive elements make controls look like controls give them strong perceived affordance use well established therefore recognizable icons and symbols assistive technology ensure content is not obscured through zooming support windows high contrast mode use images not background images provide alternative text for salient images make scrollable elements focusable for keyboard users ensure keyboard focus order is logical regarding visual layout match semantics to behavior for assistive technology users provide a default language and use lang for subsections in different languages inform the user when there are important changes to the application state do not hijack standard scrolling behavior do not instate infinite scroll by default provide buttons to load more items general accessibility make sure text and background colors contrast sufficiently do not rely on color for differentiation of visual elements avoid images of text — text that cannot be translated selected or understood by assistive tech provide a print stylesheet honour requests to remove animation via the prefers reduced motion media query seo ensure all pages have appropriate title ensure all pages have meta descriptions make content easier to find and improve search results with structured data check whether page should be appearing in sitemap make sure page has facebook and twitter large image previews set correctly check canonical links for page mark as cornerstone content performance ensure all css assets are minified and concatenated ensure all js assets are minified and concatenated ensure all images are compressed where possible remove redundant code ensure all svg assets have been optimised make sure styles and scripts are not render blocking ensure large image assets are lazy loaded other make sure all content belongs to a landmark element provide a manifest json file for identifiable homescreen entries breakpoint testing tasks desktop provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page tablet provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page mobile provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page re run automated audits tasks lighthouse run accessibility audit in lighthouse using incognito tab run performance audit in lighthouse using incognito tab run best practices audit in lighthouse using incognito tab run seo audit in lighthouse using incognito tab run pwa audit in lighthouse using incognito tab pingdom run full audit of the the page s performance in pingdom browser s console check chrome s console for errors log results of audits screenshot snapshot of the lighthouse audits upload pdf of detailed lighthouse reports provide a screenshot of any console errors
0
37,671
10,057,386,823
IssuesEvent
2019-07-22 11:30:02
haskell/haskell-ide-engine
https://api.github.com/repos/haskell/haskell-ide-engine
closed
unable to build with cabal
build
as part of #1221, I need to build `hie` with cabal, which worked fine up to now, but I am getting the following error: ``` $ cabal new-build -w ghc-8.6.5 Resolving dependencies... cabal: Could not resolve dependencies: [__0] trying: HaRe-0.8.4.1 (user goal) [__1] rejecting: HaRe:!test (constraint from config file, command line flag, or user target requires opposite flag selection) [__1] trying: HaRe:*test [__2] trying: turtle-1.5.14 (dependency of HaRe *test) [__3] trying: temporary-1.3 (dependency of turtle) [__4] next goal: cabal-helper (user goal) [__4] rejecting: cabal-helper-0.9.0.0 (conflict: temporary==1.3, cabal-helper => temporary<1.3 && >=1.2.1) [__4] rejecting: cabal-helper-0.8.2.0, cabal-helper-0.8.1.2, cabal-helper-0.8.0.2, cabal-helper-0.8.0.1, cabal-helper-0.8.0.0, cabal-helper-0.7.3.0, cabal-helper-0.7.2.0, cabal-helper-0.7.1.0, cabal-helper-0.6.3.1, cabal-helper-0.6.3.0, cabal-helper-0.6.2.0, cabal-helper-0.6.1.0, cabal-helper-0.6.0.0, cabal-helper-0.5.3.0, cabal-helper-0.5.1.0, cabal-helper-0.5.0.0, cabal-helper-0.4.0.0, cabal-helper-0.3.9.0, cabal-helper-0.3.8.0, cabal-helper-0.3.7.0, cabal-helper-0.3.6.0, cabal-helper-0.3.5.0, cabal-helper-0.3.4.0, cabal-helper-0.3.3.0, cabal-helper-0.3.2.1, cabal-helper-0.3.2.0, cabal-helper-0.3.1.0, cabal-helper-0.3.0.0, cabal-helper-0.2.0.0, cabal-helper-0.1.0.1, cabal-helper-0.1.0.0, cabal-helper-0.8.1.1, cabal-helper-0.8.1.0, cabal-helper-0.7.0.1, cabal-helper-0.5.2.0 (constraint from user target requires ==0.9.0.0) [__4] fail (backjumping, conflict set: cabal-helper, temporary) Backjump limit reached (currently 2000, change with --max-backjumps or try to run with --reorder-goals). ``` I have no problem building `hie` with stack. I am on the current master (e2f8e321bbb1bff3fc5e1d2dae649a0d61a378b1). Can anyone reproduce this issue?
1.0
unable to build with cabal - as part of #1221, I need to build `hie` with cabal, which worked fine up to now, but I am getting the following error: ``` $ cabal new-build -w ghc-8.6.5 Resolving dependencies... cabal: Could not resolve dependencies: [__0] trying: HaRe-0.8.4.1 (user goal) [__1] rejecting: HaRe:!test (constraint from config file, command line flag, or user target requires opposite flag selection) [__1] trying: HaRe:*test [__2] trying: turtle-1.5.14 (dependency of HaRe *test) [__3] trying: temporary-1.3 (dependency of turtle) [__4] next goal: cabal-helper (user goal) [__4] rejecting: cabal-helper-0.9.0.0 (conflict: temporary==1.3, cabal-helper => temporary<1.3 && >=1.2.1) [__4] rejecting: cabal-helper-0.8.2.0, cabal-helper-0.8.1.2, cabal-helper-0.8.0.2, cabal-helper-0.8.0.1, cabal-helper-0.8.0.0, cabal-helper-0.7.3.0, cabal-helper-0.7.2.0, cabal-helper-0.7.1.0, cabal-helper-0.6.3.1, cabal-helper-0.6.3.0, cabal-helper-0.6.2.0, cabal-helper-0.6.1.0, cabal-helper-0.6.0.0, cabal-helper-0.5.3.0, cabal-helper-0.5.1.0, cabal-helper-0.5.0.0, cabal-helper-0.4.0.0, cabal-helper-0.3.9.0, cabal-helper-0.3.8.0, cabal-helper-0.3.7.0, cabal-helper-0.3.6.0, cabal-helper-0.3.5.0, cabal-helper-0.3.4.0, cabal-helper-0.3.3.0, cabal-helper-0.3.2.1, cabal-helper-0.3.2.0, cabal-helper-0.3.1.0, cabal-helper-0.3.0.0, cabal-helper-0.2.0.0, cabal-helper-0.1.0.1, cabal-helper-0.1.0.0, cabal-helper-0.8.1.1, cabal-helper-0.8.1.0, cabal-helper-0.7.0.1, cabal-helper-0.5.2.0 (constraint from user target requires ==0.9.0.0) [__4] fail (backjumping, conflict set: cabal-helper, temporary) Backjump limit reached (currently 2000, change with --max-backjumps or try to run with --reorder-goals). ``` I have no problem building `hie` with stack. I am on the current master (e2f8e321bbb1bff3fc5e1d2dae649a0d61a378b1). Can anyone reproduce this issue?
non_process
unable to build with cabal as part of i need to build hie with cabal which worked fine up to now but i am getting the following error cabal new build w ghc resolving dependencies cabal could not resolve dependencies trying hare user goal rejecting hare test constraint from config file command line flag or user target requires opposite flag selection trying hare test trying turtle dependency of hare test trying temporary dependency of turtle next goal cabal helper user goal rejecting cabal helper conflict temporary cabal helper temporary rejecting cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper cabal helper constraint from user target requires fail backjumping conflict set cabal helper temporary backjump limit reached currently change with max backjumps or try to run with reorder goals i have no problem building hie with stack i am on the current master can anyone reproduce this issue
0
16,436
21,316,757,913
IssuesEvent
2022-04-16 12:16:29
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Issue with copy-to on sub-topics with locktitles
bug priority/medium preprocess preprocess2 stale
## Expected Behavior The topic should use the navtitle defined in the map when @locktitle=yes. Note: @copy-to and @locktitle works when used at the same level as the original topic. ## Actual Behavior Instead topic is using the title of the first instance of the topic when topic is referenced as a sub-topic. ## Possible Solution Fix copy-to to use the navtitle defined in the map ## Steps to Reproduce Create ditamap with topic re-used as a sub topic <map class="- map/map " domains="(map bookmap) (map mapgroup-d) (map subjectScheme) (topic abbrev-d) (topic delay-d) a(props deliveryTarget) (map ditavalref-d) (map glossref-d) (topic hazard-d) (topic hi-d) (topic indexing-d) (topic markup-d) (topic pr-d) (topic relmgmt-d) (topic sw-d) (topic ui-d) (topic ut-d) (topic markup-d xml-d) (topic xnal-d) a(loc ixia_locid)" id="axh1430149607893" title="Arron Test Map" xml:lang="en-us"> <topicref class="- map/topicref " ixia_locid="5" keys="axk1425415118966" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle " ixia_locid="7">Performing reading tasks</navtitle> <linktext class="- map/linktext " ixia_locid="8">Performing reading tasks</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> </topicref> <topicref class="- map/topicref " copy-to="second-axk1425415118966.xml" ixia_locid="11" keys="second-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle " ixia_locid="12">Arron Testing Copy-to and Locktitle</navtitle> <linktext class="- map/linktext " ixia_locid="13">Arron Testing Copy-to and Locktitle</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> <topicref class="- map/topicref " copy-to="third-axk1425415118966.xml" ixia_locid="14" keys="third-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle ">Arron 2 Testing Copy-to and Locktitle</navtitle> <linktext class="- map/linktext " ixia_locid="16">Arron 2 Testing Copy-to and Locktitle</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> </topicref> </topicref> <topicref class="- map/topicref " copy-to="fourth-axk1425415118966.xml" ixia_locid="14" keys="fourth-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle ">Arron 3 Testing Copy-to and Locktitle</navtitle> <linktext class="- map/linktext " ixia_locid="16">Arron 3 Testing Copy-to and Locktitle</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> </topicref> </map> <!-- Create a Gist via <https://gist.github.com/> to upload your test files. --> <!-- Link to the Gist from the issue or attach a .zip archive of your files. --> ## Copy of the error message, log file or stack trace None ## Environment * DITA-OT version: 3.3.4 * Operating system and version:windows * How did you run DITA-OT? oxygen * Transformation type: HTML5
2.0
Issue with copy-to on sub-topics with locktitles - ## Expected Behavior The topic should use the navtitle defined in the map when @locktitle=yes. Note: @copy-to and @locktitle works when used at the same level as the original topic. ## Actual Behavior Instead topic is using the title of the first instance of the topic when topic is referenced as a sub-topic. ## Possible Solution Fix copy-to to use the navtitle defined in the map ## Steps to Reproduce Create ditamap with topic re-used as a sub topic <map class="- map/map " domains="(map bookmap) (map mapgroup-d) (map subjectScheme) (topic abbrev-d) (topic delay-d) a(props deliveryTarget) (map ditavalref-d) (map glossref-d) (topic hazard-d) (topic hi-d) (topic indexing-d) (topic markup-d) (topic pr-d) (topic relmgmt-d) (topic sw-d) (topic ui-d) (topic ut-d) (topic markup-d xml-d) (topic xnal-d) a(loc ixia_locid)" id="axh1430149607893" title="Arron Test Map" xml:lang="en-us"> <topicref class="- map/topicref " ixia_locid="5" keys="axk1425415118966" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle " ixia_locid="7">Performing reading tasks</navtitle> <linktext class="- map/linktext " ixia_locid="8">Performing reading tasks</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> </topicref> <topicref class="- map/topicref " copy-to="second-axk1425415118966.xml" ixia_locid="11" keys="second-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle " ixia_locid="12">Arron Testing Copy-to and Locktitle</navtitle> <linktext class="- map/linktext " ixia_locid="13">Arron Testing Copy-to and Locktitle</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> <topicref class="- map/topicref " copy-to="third-axk1425415118966.xml" ixia_locid="14" keys="third-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle ">Arron 2 Testing Copy-to and Locktitle</navtitle> <linktext class="- map/linktext " ixia_locid="16">Arron 2 Testing Copy-to and Locktitle</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> </topicref> </topicref> <topicref class="- map/topicref " copy-to="fourth-axk1425415118966.xml" ixia_locid="14" keys="fourth-axk1425415118966" locktitle="yes" href="axk1425415118966.xml" type="task"> <topicmeta class="- map/topicmeta "> <navtitle class="- topic/navtitle ">Arron 3 Testing Copy-to and Locktitle</navtitle> <linktext class="- map/linktext " ixia_locid="16">Arron 3 Testing Copy-to and Locktitle</linktext> <shortdesc class="- map/shortdesc ">Whether dictating with online speech recognition, digital dictation, or typing reports, you report on studies by way of reading tasks. A single reading task can contain one or more studies for which you create a single report.</shortdesc> </topicmeta> </topicref> </map> <!-- Create a Gist via <https://gist.github.com/> to upload your test files. --> <!-- Link to the Gist from the issue or attach a .zip archive of your files. --> ## Copy of the error message, log file or stack trace None ## Environment * DITA-OT version: 3.3.4 * Operating system and version:windows * How did you run DITA-OT? oxygen * Transformation type: HTML5
process
issue with copy to on sub topics with locktitles expected behavior the topic should use the navtitle defined in the map when locktitle yes note copy to and locktitle works when used at the same level as the original topic actual behavior instead topic is using the title of the first instance of the topic when topic is referenced as a sub topic possible solution fix copy to to use the navtitle defined in the map steps to reproduce create ditamap with topic re used as a sub topic map class map map domains map bookmap map mapgroup d map subjectscheme topic abbrev d topic delay d a props deliverytarget map ditavalref d map glossref d topic hazard d topic hi d topic indexing d topic markup d topic pr d topic relmgmt d topic sw d topic ui d topic ut d topic markup d xml d topic xnal d a loc ixia locid id title arron test map xml lang en us topicref class map topicref ixia locid keys href xml type task performing reading tasks performing reading tasks whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report topicref class map topicref copy to second xml ixia locid keys second locktitle yes href xml type task arron testing copy to and locktitle arron testing copy to and locktitle whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report topicref class map topicref copy to third xml ixia locid keys third locktitle yes href xml type task arron testing copy to and locktitle arron testing copy to and locktitle whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report topicref class map topicref copy to fourth xml ixia locid keys fourth locktitle yes href xml type task arron testing copy to and locktitle arron testing copy to and locktitle whether dictating with online speech recognition digital dictation or typing reports you report on studies by way of reading tasks a single reading task can contain one or more studies for which you create a single report copy of the error message log file or stack trace none environment dita ot version operating system and version windows how did you run dita ot oxygen transformation type
1
726,607
25,004,794,859
IssuesEvent
2022-11-03 10:56:39
sapa/performing-arts-ch-templates
https://api.github.com/repos/sapa/performing-arts-ch-templates
closed
New Field : rico:hasOrHadIdentifier
Neues Feld Priority A ToValidate Blocker BR-02 BR-03 BR-04
1. Creation Properties 2. Add: Form:Record Set, Record, Instantation 3. Display : Template:Record Set, Record, Instantation ## Class of **Properties** zu erstellen - rico:hasOrHadIdentifier -- URI : https://www.ica.org/standards/RiC/ontology#hasOrHadIdentifier -- rdf:type : ?owl:ObjectProperty -- rdfs:label : has or had identifier@en - rico:isIdentifierTypeOf -- URI : https://www.ica.org/standards/RiC/ontology#isIdentifierTypeOf -- rdf.type : ?owl:ObjectProperty -- rdfs:label : is Identifier Type of @en ## Class of **Entities** zu erstellen - rico:Identifier mit properties: -- URI : https://www.ica.org/standards/RiC/ontology#Identifier -- rdf:type ?owl:Class -- rdfs:label Identifier@en - rico:IdentifierType mit proterties: -- URI : https://www.ica.org/standards/RiC/ontology#IdentifierType -- rdf:type ?owl:Class -- rdfs:label : Identifier Type@en ## Vocabulary zu erstellen - New Identifier vocabulary URI : http://vocab.performing-arts.ch/id -- Entity: **old identifier** --- rdf:type rico:IdentifierTyp --- URI : http://vocab.performing-arts.ch/idold --- rdfs:label : old identifier@en ancien identifiant@fr alt Signatur@de -- Entity : **Unique ID** --- rdfs:label : Unique ID@en Cote d'identification@fr Identifizierung Signatur@de --- URI : http://vocab.performing-arts.ch/iduni -- Entity: **technical Identifier** --- rdf:type rico:IdentifierTyp --- URI : http://vocab.performing-arts.ch/idtec --- rdfs:label : Technical Identifier@en Identifiant technique@fr (à définir)@de -- Entity: **contextual Identifier** --- rdf:type rico:IdentifierTyp --- URI : http://vocab.performing-arts.ch/idcon --- rdfs:label : Contextual Identifier@en Cote de contexte@fr (à définir)@de ## Form zu modifizieren: Record Set, Record (and RecordPart), Instantation - Hinzufügen (wie Add Daten oder Add Extent) module "Add Identifier" (add property rico:hasOrHadIdentifier and create a rico:Identifier Entity URI: http://data.performing-arts.ch/x/UUID) mit : -- Wahl zwischen vorerstellen vocabulary von rico:IdentifierType Entities (ex: Old Identifier, Cote d'individualisation)->rico:isIdentifierTypeOf -- string value ->rdf:value ## Template Display the value of rico:hasOrHadIdentifier In Record Set, Record and Instantation Template under the existing label "Identifier" with the type between brackets. ## to do after rico:identifier #166 -> #165-> #167->#168->#168 spao:legacyidentifier #170 RecordPart Order : #177
1.0
New Field : rico:hasOrHadIdentifier - 1. Creation Properties 2. Add: Form:Record Set, Record, Instantation 3. Display : Template:Record Set, Record, Instantation ## Class of **Properties** zu erstellen - rico:hasOrHadIdentifier -- URI : https://www.ica.org/standards/RiC/ontology#hasOrHadIdentifier -- rdf:type : ?owl:ObjectProperty -- rdfs:label : has or had identifier@en - rico:isIdentifierTypeOf -- URI : https://www.ica.org/standards/RiC/ontology#isIdentifierTypeOf -- rdf.type : ?owl:ObjectProperty -- rdfs:label : is Identifier Type of @en ## Class of **Entities** zu erstellen - rico:Identifier mit properties: -- URI : https://www.ica.org/standards/RiC/ontology#Identifier -- rdf:type ?owl:Class -- rdfs:label Identifier@en - rico:IdentifierType mit proterties: -- URI : https://www.ica.org/standards/RiC/ontology#IdentifierType -- rdf:type ?owl:Class -- rdfs:label : Identifier Type@en ## Vocabulary zu erstellen - New Identifier vocabulary URI : http://vocab.performing-arts.ch/id -- Entity: **old identifier** --- rdf:type rico:IdentifierTyp --- URI : http://vocab.performing-arts.ch/idold --- rdfs:label : old identifier@en ancien identifiant@fr alt Signatur@de -- Entity : **Unique ID** --- rdfs:label : Unique ID@en Cote d'identification@fr Identifizierung Signatur@de --- URI : http://vocab.performing-arts.ch/iduni -- Entity: **technical Identifier** --- rdf:type rico:IdentifierTyp --- URI : http://vocab.performing-arts.ch/idtec --- rdfs:label : Technical Identifier@en Identifiant technique@fr (à définir)@de -- Entity: **contextual Identifier** --- rdf:type rico:IdentifierTyp --- URI : http://vocab.performing-arts.ch/idcon --- rdfs:label : Contextual Identifier@en Cote de contexte@fr (à définir)@de ## Form zu modifizieren: Record Set, Record (and RecordPart), Instantation - Hinzufügen (wie Add Daten oder Add Extent) module "Add Identifier" (add property rico:hasOrHadIdentifier and create a rico:Identifier Entity URI: http://data.performing-arts.ch/x/UUID) mit : -- Wahl zwischen vorerstellen vocabulary von rico:IdentifierType Entities (ex: Old Identifier, Cote d'individualisation)->rico:isIdentifierTypeOf -- string value ->rdf:value ## Template Display the value of rico:hasOrHadIdentifier In Record Set, Record and Instantation Template under the existing label "Identifier" with the type between brackets. ## to do after rico:identifier #166 -> #165-> #167->#168->#168 spao:legacyidentifier #170 RecordPart Order : #177
non_process
new field rico hasorhadidentifier creation properties add form record set record instantation display template record set record instantation class of properties zu erstellen rico hasorhadidentifier uri rdf type owl objectproperty rdfs label has or had identifier en rico isidentifiertypeof uri rdf type owl objectproperty rdfs label is identifier type of en class of entities zu erstellen rico identifier mit properties uri rdf type owl class rdfs label identifier en rico identifiertype mit proterties uri rdf type owl class rdfs label identifier type en vocabulary zu erstellen new identifier vocabulary uri entity old identifier rdf type rico identifiertyp uri rdfs label old identifier en ancien identifiant fr alt signatur de entity unique id rdfs label unique id en cote d identification fr identifizierung signatur de uri entity technical identifier rdf type rico identifiertyp uri rdfs label technical identifier en identifiant technique fr à définir de entity contextual identifier rdf type rico identifiertyp uri rdfs label contextual identifier en cote de contexte fr à définir de form zu modifizieren record set record and recordpart instantation hinzufügen wie add daten oder add extent module add identifier add property rico hasorhadidentifier and create a rico identifier entity uri mit wahl zwischen vorerstellen vocabulary von rico identifiertype entities ex old identifier cote d individualisation rico isidentifiertypeof string value rdf value template display the value of rico hasorhadidentifier in record set record and instantation template under the existing label identifier with the type between brackets to do after rico identifier spao legacyidentifier recordpart order
0
13,838
16,600,558,174
IssuesEvent
2021-06-01 18:49:09
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
closed
Add proper argument typing to `sql` components
annoying bughancement inputs processors
Hi, I found that in the sql plugin, it converts all data to string and then exec the sql query. However, for clickhouse it will report an error. Do you think it is possible to fix this issue?
1.0
Add proper argument typing to `sql` components - Hi, I found that in the sql plugin, it converts all data to string and then exec the sql query. However, for clickhouse it will report an error. Do you think it is possible to fix this issue?
process
add proper argument typing to sql components hi i found that in the sql plugin it converts all data to string and then exec the sql query however for clickhouse it will report an error do you think it is possible to fix this issue
1
1,711
4,351,016,870
IssuesEvent
2016-07-31 16:20:56
spootTheLousy/saguaro
https://api.github.com/repos/spootTheLousy/saguaro
opened
Check images for hidden archives
Image processing Security
Currently, saguaro does no sanitation or checks in regards to images that have hidden embedded content. More often than not, it's a method of transmitting illegal content, so it is definitely within the scope of the repo to check images for this. Fortunately, PHP 5.2 has two classes that ship with it for handling `.rar` and `.zip` files, should be as simple as checking if an image is extractable, and if it is, rejecting it.
1.0
Check images for hidden archives - Currently, saguaro does no sanitation or checks in regards to images that have hidden embedded content. More often than not, it's a method of transmitting illegal content, so it is definitely within the scope of the repo to check images for this. Fortunately, PHP 5.2 has two classes that ship with it for handling `.rar` and `.zip` files, should be as simple as checking if an image is extractable, and if it is, rejecting it.
process
check images for hidden archives currently saguaro does no sanitation or checks in regards to images that have hidden embedded content more often than not it s a method of transmitting illegal content so it is definitely within the scope of the repo to check images for this fortunately php has two classes that ship with it for handling rar and zip files should be as simple as checking if an image is extractable and if it is rejecting it
1
16,620
9,853,207,050
IssuesEvent
2019-06-19 14:21:35
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Object Level Security
Team:Security enhancement
## Object Level Security ### ACL To enable OLS, an ACL with the following format will be attached to all securable objects: ``` { "owner": 123456789, "read": { "users": [ { id: 123456789, can_share: true } ], "roles": [ { id: "role_one", can_share: false } ] }, "write": { "users": [ { id: 123456789, can_share: true } ], "roles": [ { id: "role_one", can_share: false } ] } } ``` The users and roles above reference Elasticsearch users and roles. In addition to the Elasticsearch roles, we will utilize a role of `*` to denote all authenticated users. If the user has **write** they will implicitly be granted **read**. When a user or role is assigned **read** or **write** they will be able to specify whether this user or role will be able to share the securable object with others. If the user has **read** and they can share the object, they will only be able to add other users and roles to **read**. If the user has **write** and they can share the object, they will be able to add other users and roles to **write** and **read**. ### Implicit read permissions When a user has **read** access to a Dashboard, they will implicitly be granted **read** access to all related Visualizations and Saved Searches. The same logic will apply once Index Patterns themselves are made securable and if a user has **read** access to a Visualization or Saved Search, they will be implicitly granted **read** access to the Index Pattern. This simplifies the access model and allows users to assign access to the object that they intuitively wish to share without having to concern themselves with the graph of related objects. It also simplifies the technical implementation so we don’t have to explicitly assign access to the related objects and then determine if/when it should be removed when a parent object’s ACL is modified. When a user is implicitly granted **read** access to a Visualization or Saved Search, it won’t show up in the user’s list of Visualizations or Saved Searches, it will only be accessible in the Dashboard UI/API. This is similar to how we’ll implement it technically, we’ll allow users to gain access to the related objects via the Dashboard, which will implicitly be granting them **read** access. ### Summary Phase 4 will make Saved Searches, Dashboards, Visualizations, Index Patterns and other Kibana applications (Machine Learning, Graph, Timelion) saved objects securable based on the previously described ACL. When an object has no owner, it emulates the way that Kibana currently functions without OLS where all authenticated Kibana users have full permissions. This is purely to support migrations from older versions of Kibana that didn’t have OLS, or users that were running Kibana without security and then enabling security with OLS. An additional “Claim unowned object” privilege will be added to the kibana_user role, and the user will have to have this privilege to claim these unowned objects. The introduction of owned Index Patterns necessitates the addition of per-user Kibana Advanced Settings, as the default index pattern is defined here. An additional section will be added to the Advanced Settings page to allow a user to override any advanced setting, the same capability will be added to the index management page. When a securable object has no owner, they will see a dialog similar to the following allowing them to make themselves the owner: ![screen shot 2018-03-09 at 10 28 06 am](https://user-images.githubusercontent.com/627123/37424187-4d01f5f0-2796-11e8-9e8d-cc2608f420a8.png) A securable object with no owner will be represented by the non-existence of an ACL. When a securable object has an owner, they will see a dialog similar to the following allowing them to transfer ownership and define which users and roles can read/write the object: ![untitled](https://user-images.githubusercontent.com/627123/37487123-f870d3b8-2866-11e8-94c8-4cff349c810c.png) System administrators will always be able to transfer ownership amd modify the ACL of a securable object, incase a user erroneously claims ownership of an owned object. All users that have a role granting them a Kibana custom privilege for the specific Kibana instance will be listed, and all roles that have a Kibana custom privilege for the Kibana specific instance will be listed as well. It should be noted that for Kibana to be able to fully enumerate users, we will have to introduce the concept of user profiles in Kibana (that could potentially power the user specific settings) or have Elasticsearch create users for non-native realms. Currently, Elasticsearch is unable to enumerate all users for SAML/LDAP/etc. realms as these are powered by role mappings. The list of Saved Searches, Dashboards, Visualizations and Index Patterns will have an owner column added, similar to the following: ![screen shot 2018-03-09 at 11 11 16 am](https://user-images.githubusercontent.com/627123/37424269-7553c060-2796-11e8-9d4f-3094c33469cd.png) From this phase forward, all new securable objects will be owned by the creator and they will have to share them with others. This same logic applies to objects that are imported. They will be owned by the user importing them, and can then be shared. Additional Kibana applications (Graph, Timelion) will be modified to support a similar mechanism of claiming/transferring ownership, and listing the current owner. In the future, there’s potential for the Kibana admin to be able to define default permissions for different users, or to use RBAC to limit users being able to create private or public securable objects. However, this level of control will not be introduced in this phase, as it might not be needed and it increases the complexity and implementation time.
True
Object Level Security - ## Object Level Security ### ACL To enable OLS, an ACL with the following format will be attached to all securable objects: ``` { "owner": 123456789, "read": { "users": [ { id: 123456789, can_share: true } ], "roles": [ { id: "role_one", can_share: false } ] }, "write": { "users": [ { id: 123456789, can_share: true } ], "roles": [ { id: "role_one", can_share: false } ] } } ``` The users and roles above reference Elasticsearch users and roles. In addition to the Elasticsearch roles, we will utilize a role of `*` to denote all authenticated users. If the user has **write** they will implicitly be granted **read**. When a user or role is assigned **read** or **write** they will be able to specify whether this user or role will be able to share the securable object with others. If the user has **read** and they can share the object, they will only be able to add other users and roles to **read**. If the user has **write** and they can share the object, they will be able to add other users and roles to **write** and **read**. ### Implicit read permissions When a user has **read** access to a Dashboard, they will implicitly be granted **read** access to all related Visualizations and Saved Searches. The same logic will apply once Index Patterns themselves are made securable and if a user has **read** access to a Visualization or Saved Search, they will be implicitly granted **read** access to the Index Pattern. This simplifies the access model and allows users to assign access to the object that they intuitively wish to share without having to concern themselves with the graph of related objects. It also simplifies the technical implementation so we don’t have to explicitly assign access to the related objects and then determine if/when it should be removed when a parent object’s ACL is modified. When a user is implicitly granted **read** access to a Visualization or Saved Search, it won’t show up in the user’s list of Visualizations or Saved Searches, it will only be accessible in the Dashboard UI/API. This is similar to how we’ll implement it technically, we’ll allow users to gain access to the related objects via the Dashboard, which will implicitly be granting them **read** access. ### Summary Phase 4 will make Saved Searches, Dashboards, Visualizations, Index Patterns and other Kibana applications (Machine Learning, Graph, Timelion) saved objects securable based on the previously described ACL. When an object has no owner, it emulates the way that Kibana currently functions without OLS where all authenticated Kibana users have full permissions. This is purely to support migrations from older versions of Kibana that didn’t have OLS, or users that were running Kibana without security and then enabling security with OLS. An additional “Claim unowned object” privilege will be added to the kibana_user role, and the user will have to have this privilege to claim these unowned objects. The introduction of owned Index Patterns necessitates the addition of per-user Kibana Advanced Settings, as the default index pattern is defined here. An additional section will be added to the Advanced Settings page to allow a user to override any advanced setting, the same capability will be added to the index management page. When a securable object has no owner, they will see a dialog similar to the following allowing them to make themselves the owner: ![screen shot 2018-03-09 at 10 28 06 am](https://user-images.githubusercontent.com/627123/37424187-4d01f5f0-2796-11e8-9e8d-cc2608f420a8.png) A securable object with no owner will be represented by the non-existence of an ACL. When a securable object has an owner, they will see a dialog similar to the following allowing them to transfer ownership and define which users and roles can read/write the object: ![untitled](https://user-images.githubusercontent.com/627123/37487123-f870d3b8-2866-11e8-94c8-4cff349c810c.png) System administrators will always be able to transfer ownership amd modify the ACL of a securable object, incase a user erroneously claims ownership of an owned object. All users that have a role granting them a Kibana custom privilege for the specific Kibana instance will be listed, and all roles that have a Kibana custom privilege for the Kibana specific instance will be listed as well. It should be noted that for Kibana to be able to fully enumerate users, we will have to introduce the concept of user profiles in Kibana (that could potentially power the user specific settings) or have Elasticsearch create users for non-native realms. Currently, Elasticsearch is unable to enumerate all users for SAML/LDAP/etc. realms as these are powered by role mappings. The list of Saved Searches, Dashboards, Visualizations and Index Patterns will have an owner column added, similar to the following: ![screen shot 2018-03-09 at 11 11 16 am](https://user-images.githubusercontent.com/627123/37424269-7553c060-2796-11e8-9d4f-3094c33469cd.png) From this phase forward, all new securable objects will be owned by the creator and they will have to share them with others. This same logic applies to objects that are imported. They will be owned by the user importing them, and can then be shared. Additional Kibana applications (Graph, Timelion) will be modified to support a similar mechanism of claiming/transferring ownership, and listing the current owner. In the future, there’s potential for the Kibana admin to be able to define default permissions for different users, or to use RBAC to limit users being able to create private or public securable objects. However, this level of control will not be introduced in this phase, as it might not be needed and it increases the complexity and implementation time.
non_process
object level security object level security acl to enable ols an acl with the following format will be attached to all securable objects owner read users roles write users roles the users and roles above reference elasticsearch users and roles in addition to the elasticsearch roles we will utilize a role of to denote all authenticated users if the user has write they will implicitly be granted read when a user or role is assigned read or write they will be able to specify whether this user or role will be able to share the securable object with others if the user has read and they can share the object they will only be able to add other users and roles to read if the user has write and they can share the object they will be able to add other users and roles to write and read implicit read permissions when a user has read access to a dashboard they will implicitly be granted read access to all related visualizations and saved searches the same logic will apply once index patterns themselves are made securable and if a user has read access to a visualization or saved search they will be implicitly granted read access to the index pattern this simplifies the access model and allows users to assign access to the object that they intuitively wish to share without having to concern themselves with the graph of related objects it also simplifies the technical implementation so we don’t have to explicitly assign access to the related objects and then determine if when it should be removed when a parent object’s acl is modified when a user is implicitly granted read access to a visualization or saved search it won’t show up in the user’s list of visualizations or saved searches it will only be accessible in the dashboard ui api this is similar to how we’ll implement it technically we’ll allow users to gain access to the related objects via the dashboard which will implicitly be granting them read access summary phase will make saved searches dashboards visualizations index patterns and other kibana applications machine learning graph timelion saved objects securable based on the previously described acl when an object has no owner it emulates the way that kibana currently functions without ols where all authenticated kibana users have full permissions this is purely to support migrations from older versions of kibana that didn’t have ols or users that were running kibana without security and then enabling security with ols an additional “claim unowned object” privilege will be added to the kibana user role and the user will have to have this privilege to claim these unowned objects the introduction of owned index patterns necessitates the addition of per user kibana advanced settings as the default index pattern is defined here an additional section will be added to the advanced settings page to allow a user to override any advanced setting the same capability will be added to the index management page when a securable object has no owner they will see a dialog similar to the following allowing them to make themselves the owner a securable object with no owner will be represented by the non existence of an acl when a securable object has an owner they will see a dialog similar to the following allowing them to transfer ownership and define which users and roles can read write the object system administrators will always be able to transfer ownership amd modify the acl of a securable object incase a user erroneously claims ownership of an owned object all users that have a role granting them a kibana custom privilege for the specific kibana instance will be listed and all roles that have a kibana custom privilege for the kibana specific instance will be listed as well it should be noted that for kibana to be able to fully enumerate users we will have to introduce the concept of user profiles in kibana that could potentially power the user specific settings or have elasticsearch create users for non native realms currently elasticsearch is unable to enumerate all users for saml ldap etc realms as these are powered by role mappings the list of saved searches dashboards visualizations and index patterns will have an owner column added similar to the following from this phase forward all new securable objects will be owned by the creator and they will have to share them with others this same logic applies to objects that are imported they will be owned by the user importing them and can then be shared additional kibana applications graph timelion will be modified to support a similar mechanism of claiming transferring ownership and listing the current owner in the future there’s potential for the kibana admin to be able to define default permissions for different users or to use rbac to limit users being able to create private or public securable objects however this level of control will not be introduced in this phase as it might not be needed and it increases the complexity and implementation time
0
6,466
9,546,611,790
IssuesEvent
2019-05-01 20:27:31
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Department of State Drawers
Apply Process Requirements Ready State Dept.
Who: Student What: ability to edit or delete information in Open Opps Why: As a student applying for an internship I would like the ability to edit and delete information from my application A/C - The drawers throughout the application will follow the design system https://usajobs.github.io/design-system/drawer/ - Each drawer will for References, Experience and Languages will have an edit button that will allow a user to edit the information in Open Opportunities only - Each drawer for References, Experience and Languages will have a delete button that will allow the user to delete the information in Open Opportunities only
1.0
Department of State Drawers - Who: Student What: ability to edit or delete information in Open Opps Why: As a student applying for an internship I would like the ability to edit and delete information from my application A/C - The drawers throughout the application will follow the design system https://usajobs.github.io/design-system/drawer/ - Each drawer will for References, Experience and Languages will have an edit button that will allow a user to edit the information in Open Opportunities only - Each drawer for References, Experience and Languages will have a delete button that will allow the user to delete the information in Open Opportunities only
process
department of state drawers who student what ability to edit or delete information in open opps why as a student applying for an internship i would like the ability to edit and delete information from my application a c the drawers throughout the application will follow the design system each drawer will for references experience and languages will have an edit button that will allow a user to edit the information in open opportunities only each drawer for references experience and languages will have a delete button that will allow the user to delete the information in open opportunities only
1
121,162
10,152,085,424
IssuesEvent
2019-08-05 22:15:47
MadalinNitu/Bendis
https://api.github.com/repos/MadalinNitu/Bendis
closed
Update Profile models for allow guest accounts.
back-end medium testing
Update Profile models for allow guest accounts.
1.0
Update Profile models for allow guest accounts. - Update Profile models for allow guest accounts.
non_process
update profile models for allow guest accounts update profile models for allow guest accounts
0
99,720
4,063,715,432
IssuesEvent
2016-05-26 01:24:22
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Problem with visibility
Category: Entities Category: Miscellaneous Priority: High Type: Bug
When you delete a creature, it does not dissapear until you move. May be related to #1199
1.0
Problem with visibility - When you delete a creature, it does not dissapear until you move. May be related to #1199
non_process
problem with visibility when you delete a creature it does not dissapear until you move may be related to
0
19,776
26,156,392,637
IssuesEvent
2022-12-30 22:48:19
kitspace/kitspace-v2
https://api.github.com/repos/kitspace/kitspace-v2
opened
Docker volumes size increases significantly after importing all projects
processor
I have noticed that the deployment to my staging server has been starting to fail frequently due to disk space. Checking the disk usage I have found that the processor and gitea volumes are taking up ~**25GB** which causes image pull operations to fail. We can either: 1. upgrade the server, 2. delete the repos from the processor in a clean-up job. I prefer the second option. ## ```console $ docker system df -v Images space usage: REPOSITORY TAG IMAGE ID CREATED SIZE SHARED SIZE UNIQUE SIZE CONTAINERS Containers space usage: CONTAINER ID IMAGE COMMAND LOCAL VOLUMES SIZE CREATED STATUS NAMES Local Volumes space usage: VOLUME NAME LINKS SIZE kitspace-abdo-dev_gitea-data 0 8.128GB <--------- fb020e97396eef361bf914971fca183126044a8554e38dd1858e223cd8428716 0 520.1kB kitspace-abdo-dev_postgres-data 0 76.63MB kitspace-abdo-dev_processor-data 0 16.84GB <--------- kitspace-abdo-dev_meilisearch-data 0 71.21MB 9cb3800976bc29e7db7bb37e8f98f70322d695d0fa2994f9efdf4c1cce18e328 0 21.27kB ```
1.0
Docker volumes size increases significantly after importing all projects - I have noticed that the deployment to my staging server has been starting to fail frequently due to disk space. Checking the disk usage I have found that the processor and gitea volumes are taking up ~**25GB** which causes image pull operations to fail. We can either: 1. upgrade the server, 2. delete the repos from the processor in a clean-up job. I prefer the second option. ## ```console $ docker system df -v Images space usage: REPOSITORY TAG IMAGE ID CREATED SIZE SHARED SIZE UNIQUE SIZE CONTAINERS Containers space usage: CONTAINER ID IMAGE COMMAND LOCAL VOLUMES SIZE CREATED STATUS NAMES Local Volumes space usage: VOLUME NAME LINKS SIZE kitspace-abdo-dev_gitea-data 0 8.128GB <--------- fb020e97396eef361bf914971fca183126044a8554e38dd1858e223cd8428716 0 520.1kB kitspace-abdo-dev_postgres-data 0 76.63MB kitspace-abdo-dev_processor-data 0 16.84GB <--------- kitspace-abdo-dev_meilisearch-data 0 71.21MB 9cb3800976bc29e7db7bb37e8f98f70322d695d0fa2994f9efdf4c1cce18e328 0 21.27kB ```
process
docker volumes size increases significantly after importing all projects i have noticed that the deployment to my staging server has been starting to fail frequently due to disk space checking the disk usage i have found that the processor and gitea volumes are taking up which causes image pull operations to fail we can either upgrade the server delete the repos from the processor in a clean up job i prefer the second option console docker system df v images space usage repository tag image id created size shared size unique size containers containers space usage container id image command local volumes size created status names local volumes space usage volume name links size kitspace abdo dev gitea data kitspace abdo dev postgres data kitspace abdo dev processor data kitspace abdo dev meilisearch data
1
15,557
19,703,503,472
IssuesEvent
2022-01-12 19:08:00
googleapis/java-securitycenter-settings
https://api.github.com/repos/googleapis/java-securitycenter-settings
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'securitycenter-settings' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'securitycenter-settings' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname securitycenter settings invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
688,526
23,586,433,079
IssuesEvent
2022-08-23 12:02:17
mm-ninja-turtles/turtle-express
https://api.github.com/repos/mm-ninja-turtles/turtle-express
closed
[bug]: in request validation
type: bug priority: high task: todo
I tested this kind of example with request body ``` const users = router.path('/users') users.handler({ id: 'CreateNewUser', method: 'post', request: { body: z.object({ name: z.string().min(1, { message: 'name is required.' }), email: z.string().email({ message: 'wrong email format.' }), }), }, response: { 200: z.object({ name: z.string(), email: z.string().email(), }), }, async resolver({ ctx }) { const { name, email } = ctx.body const user = await db.users.create({ name, email }) return { 200: user, } }, }) '''
1.0
[bug]: in request validation - I tested this kind of example with request body ``` const users = router.path('/users') users.handler({ id: 'CreateNewUser', method: 'post', request: { body: z.object({ name: z.string().min(1, { message: 'name is required.' }), email: z.string().email({ message: 'wrong email format.' }), }), }, response: { 200: z.object({ name: z.string(), email: z.string().email(), }), }, async resolver({ ctx }) { const { name, email } = ctx.body const user = await db.users.create({ name, email }) return { 200: user, } }, }) '''
non_process
in request validation i tested this kind of example with request body const users router path users users handler id createnewuser method post request body z object name z string min message name is required email z string email message wrong email format response z object name z string email z string email async resolver ctx const name email ctx body const user await db users create name email return user
0
27,197
6,819,828,528
IssuesEvent
2017-11-07 11:39:28
kimushu/rubic-vscode
https://api.github.com/repos/kimushu/rubic-vscode
closed
VSCode is no longer supporting specifiying a commandId in the "initialConfigurations" contribution
vscode
Hi, We noticed your extension in `package.json` is using a `"initialConfigurations"` contribution in a deprecated way by specifiying a command id. We wanted to let you know that we plan to delete support for this specific use of "initialConfigurations" soon. More about this deprecation can be found [here](https://github.com/Microsoft/vscode/issues/33791) and in our release notes. A good example on how to use the `DebugConfigurarationProvider` which is the new way of doing this can be found [here](https://github.com/Microsoft/vscode-node-debug/blob/d61ac00e16069fe63a5a5b6c65a4040de013e7f2/src/node/extension/configurationProvider.ts#L18) Also note other deprecations which are specified in our release notes (which can be found on our site). Kind regards, Isidor from the VSCode team
1.0
VSCode is no longer supporting specifiying a commandId in the "initialConfigurations" contribution - Hi, We noticed your extension in `package.json` is using a `"initialConfigurations"` contribution in a deprecated way by specifiying a command id. We wanted to let you know that we plan to delete support for this specific use of "initialConfigurations" soon. More about this deprecation can be found [here](https://github.com/Microsoft/vscode/issues/33791) and in our release notes. A good example on how to use the `DebugConfigurarationProvider` which is the new way of doing this can be found [here](https://github.com/Microsoft/vscode-node-debug/blob/d61ac00e16069fe63a5a5b6c65a4040de013e7f2/src/node/extension/configurationProvider.ts#L18) Also note other deprecations which are specified in our release notes (which can be found on our site). Kind regards, Isidor from the VSCode team
non_process
vscode is no longer supporting specifiying a commandid in the initialconfigurations contribution hi we noticed your extension in package json is using a initialconfigurations contribution in a deprecated way by specifiying a command id we wanted to let you know that we plan to delete support for this specific use of initialconfigurations soon more about this deprecation can be found and in our release notes a good example on how to use the debugconfigurarationprovider which is the new way of doing this can be found also note other deprecations which are specified in our release notes which can be found on our site kind regards isidor from the vscode team
0
15,992
20,188,203,910
IssuesEvent
2022-02-11 01:17:45
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Standardize on modern authentication protocols
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Authentication and authorization
<a href="https://docs.microsoft.com/azure/architecture/framework/security/design-identity-authentication#use-modern-password-protection">Standardize on modern authentication protocols</a> <p><b>Why Consider This?</b></p> Modern authentication protocols support strong controls such as MFA and should be used instead of legacy authentication methods. Use of legacy method increases risk of credential exposure. <p><b>Context</b></p> <p><span>Modern Authentication is a method of identity management that offers more secure user authentication and authorization. It's available for Office 365 hybrid deployments of Skype for Business server on-premises and Exchange server on-premises, as well as, split-domain Skype for Business hybrids. This article links to related docs about prerequisites, setup/disabling modern authentication, and to some of the related client (ex. Outlook and Skype clients) information.</span></p><p><span>Modern authentication is an umbrella term for a combination of authentication and authorization methods between a client (for example, your laptop or your phone) and a server, as well as some security measures that rely on access policies that you may already be familiar with. It includes:</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span><b>Authentication methods</b></span><span>: Multi-factor authentication (MFA); smart card authentication; client certificate-based authentication</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span><b>Authorization methods</b></span><span>: Microsoft's implementation of Open Authorization (OAuth)</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span><b>Conditional access policies</b></span><span>: Mobile Application Management (MAM) and Azure Active Directory (Azure AD) Conditional Access</span></li></ul> <p><b>Suggested Actions</b></p> <p><span>Review workloads that do not leverage modern authentication protocols and convert where possible. In addition, standarize using modern authentication protocols for all future workloads.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/microsoft-365/enterprise/hybrid-modern-auth-overview?view=o365-worldwide#BKMK_WhatisModAuth" target="_blank"><span>What is modern authentication?</span></a><span /></p>
1.0
Standardize on modern authentication protocols - <a href="https://docs.microsoft.com/azure/architecture/framework/security/design-identity-authentication#use-modern-password-protection">Standardize on modern authentication protocols</a> <p><b>Why Consider This?</b></p> Modern authentication protocols support strong controls such as MFA and should be used instead of legacy authentication methods. Use of legacy method increases risk of credential exposure. <p><b>Context</b></p> <p><span>Modern Authentication is a method of identity management that offers more secure user authentication and authorization. It's available for Office 365 hybrid deployments of Skype for Business server on-premises and Exchange server on-premises, as well as, split-domain Skype for Business hybrids. This article links to related docs about prerequisites, setup/disabling modern authentication, and to some of the related client (ex. Outlook and Skype clients) information.</span></p><p><span>Modern authentication is an umbrella term for a combination of authentication and authorization methods between a client (for example, your laptop or your phone) and a server, as well as some security measures that rely on access policies that you may already be familiar with. It includes:</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span><b>Authentication methods</b></span><span>: Multi-factor authentication (MFA); smart card authentication; client certificate-based authentication</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span><b>Authorization methods</b></span><span>: Microsoft's implementation of Open Authorization (OAuth)</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span><b>Conditional access policies</b></span><span>: Mobile Application Management (MAM) and Azure Active Directory (Azure AD) Conditional Access</span></li></ul> <p><b>Suggested Actions</b></p> <p><span>Review workloads that do not leverage modern authentication protocols and convert where possible. In addition, standarize using modern authentication protocols for all future workloads.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/microsoft-365/enterprise/hybrid-modern-auth-overview?view=o365-worldwide#BKMK_WhatisModAuth" target="_blank"><span>What is modern authentication?</span></a><span /></p>
process
standardize on modern authentication protocols why consider this modern authentication protocols support strong controls such as mfa and should be used instead of legacy authentication methods use of legacy method increases risk of credential exposure context modern authentication is a method of identity management that offers more secure user authentication and authorization it s available for office hybrid deployments of skype for business server on premises and exchange server on premises as well as split domain skype for business hybrids this article links to related docs about prerequisites setup disabling modern authentication and to some of the related client ex outlook and skype clients information modern authentication is an umbrella term for a combination of authentication and authorization methods between a client for example your laptop or your phone and a server as well as some security measures that rely on access policies that you may already be familiar with it includes authentication methods multi factor authentication mfa smart card authentication client certificate based authentication authorization methods microsoft s implementation of open authorization oauth conditional access policies mobile application management mam and azure active directory azure ad conditional access suggested actions review workloads that do not leverage modern authentication protocols and convert where possible in addition standarize using modern authentication protocols for all future workloads learn more what is modern authentication
1
2,604
5,357,060,864
IssuesEvent
2017-02-20 17:14:37
PagerNation/PagerNation
https://api.github.com/repos/PagerNation/PagerNation
closed
Read UI notes & adjust
process
## Backlog ### Problem > Give a description of the card. What is the problem you want to solve? ### Why? > What will completion of this card gain us? > Link back to a Use Case or Functional Requirement that this will solve ## Elaboration ### How do we solve this? > Get as specific as possible, talk about specific files, endpoints, etc. Anyone should be able to pick up the card and complete it in basically the same way as anyone else ### When do we know this is done? > Outline acceptance criteria ### Tests!?!!?! > Describe the tests that need to be implemented. If it's a complicated feature, get descriptive. Otherwise it's fine to be less verbose.
1.0
Read UI notes & adjust - ## Backlog ### Problem > Give a description of the card. What is the problem you want to solve? ### Why? > What will completion of this card gain us? > Link back to a Use Case or Functional Requirement that this will solve ## Elaboration ### How do we solve this? > Get as specific as possible, talk about specific files, endpoints, etc. Anyone should be able to pick up the card and complete it in basically the same way as anyone else ### When do we know this is done? > Outline acceptance criteria ### Tests!?!!?! > Describe the tests that need to be implemented. If it's a complicated feature, get descriptive. Otherwise it's fine to be less verbose.
process
read ui notes adjust backlog problem give a description of the card what is the problem you want to solve why what will completion of this card gain us link back to a use case or functional requirement that this will solve elaboration how do we solve this get as specific as possible talk about specific files endpoints etc anyone should be able to pick up the card and complete it in basically the same way as anyone else when do we know this is done outline acceptance criteria tests describe the tests that need to be implemented if it s a complicated feature get descriptive otherwise it s fine to be less verbose
1
10,746
13,540,496,793
IssuesEvent
2020-09-16 14:44:07
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
opened
Add badges to main page/readme
process + tools
We should consider adding badges for: * master build * master doc publish * future validation suite
1.0
Add badges to main page/readme - We should consider adding badges for: * master build * master doc publish * future validation suite
process
add badges to main page readme we should consider adding badges for master build master doc publish future validation suite
1
144,000
11,590,804,873
IssuesEvent
2020-02-24 07:55:13
INTO-CPS-Association/pyfmu
https://api.github.com/repos/INTO-CPS-Association/pyfmu
opened
functional testing of example projects
testing
Currently, the test suite only tests the functionality of the exported example projects by checking if the simulation is successful. A good addition would be to add tests that trace of the simulation to ensure they behave correctly.
1.0
functional testing of example projects - Currently, the test suite only tests the functionality of the exported example projects by checking if the simulation is successful. A good addition would be to add tests that trace of the simulation to ensure they behave correctly.
non_process
functional testing of example projects currently the test suite only tests the functionality of the exported example projects by checking if the simulation is successful a good addition would be to add tests that trace of the simulation to ensure they behave correctly
0
264,703
23,134,419,964
IssuesEvent
2022-07-28 13:16:16
etcd-io/etcd
https://api.github.com/repos/etcd-io/etcd
closed
FLAKE: TestKVDelete
Help Wanted area/testing/flake
https://github.com/etcd-io/etcd/runs/7499725943?check_suite_focus=true Interesting log: ``` /home/runner/work/etcd/etcd/bin/etcd-25280: {"level":"error","ts":"2022-07-25T12:53:51.450Z","caller":"embed/etcd.go:778","msg":"setting up serving from embedded etcd failed.","error":"accept tcp 127.0.0.1:20000: use of closed network connection","stacktrace":"go.etcd.io/etcd/server/v3/embed.(*Etcd).errHandler\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:778\ngo.etcd.io/etcd/server/v3/embed.(*Etcd).serveClients.func1\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:735"} ```
1.0
FLAKE: TestKVDelete - https://github.com/etcd-io/etcd/runs/7499725943?check_suite_focus=true Interesting log: ``` /home/runner/work/etcd/etcd/bin/etcd-25280: {"level":"error","ts":"2022-07-25T12:53:51.450Z","caller":"embed/etcd.go:778","msg":"setting up serving from embedded etcd failed.","error":"accept tcp 127.0.0.1:20000: use of closed network connection","stacktrace":"go.etcd.io/etcd/server/v3/embed.(*Etcd).errHandler\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:778\ngo.etcd.io/etcd/server/v3/embed.(*Etcd).serveClients.func1\n\t/home/runner/work/etcd/etcd/server/embed/etcd.go:735"} ```
non_process
flake testkvdelete interesting log home runner work etcd etcd bin etcd level error ts caller embed etcd go msg setting up serving from embedded etcd failed error accept tcp use of closed network connection stacktrace go etcd io etcd server embed etcd errhandler n t home runner work etcd etcd server embed etcd go ngo etcd io etcd server embed etcd serveclients n t home runner work etcd etcd server embed etcd go
0
350,989
31,932,708,394
IssuesEvent
2023-09-19 08:30:30
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix tensor.test_tensorflow_tensor_shape
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6201081410/job/16837044459"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix tensor.test_tensorflow_tensor_shape - | | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6201081410/job/16837044459"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6219910913"><img src=https://img.shields.io/badge/-failure-red></a>
non_process
fix tensor test tensorflow tensor shape numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
0
16,366
21,053,352,179
IssuesEvent
2022-03-31 22:53:16
googleapis/google-auth-library-ruby
https://api.github.com/repos/googleapis/google-auth-library-ruby
closed
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * could not parse .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * could not parse .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 could not parse repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
181,382
6,659,216,936
IssuesEvent
2017-10-01 08:09:12
k0shk0sh/FastHub
https://api.github.com/repos/k0shk0sh/FastHub
closed
Annoying and blocking "Too many changes to show"
Priority: Critical Status: Accepted Type: Enhancement
**FastHub Version: 4.3.0** **Android Version: 7.0 (SDK: 24)** **Device Information:** - **Manufacturer:** samsung - **Brand:** samsung - **Model:** Galaxy Tab S2 8.0 --- It seems that you added a threshold in a new version and don't show a diff in PR if file contains many changes. This is a horrible decision that significantly decrease the usability. First of all this threshold seems to be very small as most of changes are not being displayed more. Another point is that PR review feature is affected. You sell a PRO feature to perform a code review via app, however how could it be possible to left a comment if diff isn't displayed? Now I'm able to left comments to a few files in only, while for rest I should use the browser. I'm very disappointed by this bug :pensive: For me it's about the current approach in general - in each version you introduce some new "features", that later create the usability issues (minor or major) :cry: It's a bit annoying as I feel like a beta tester.. Of course, you quickly fix them, but for couple of weeks we should live with them :( Could you please consider adding an advanced configuration for such things, so we could quickly disable/tune new features if there issues with them? For instance, it would be enough to tune the threshold value to work around the current issue, rather than be unable to perform a code review for a few weeks.. Thank you for your time and work on this project!
1.0
Annoying and blocking "Too many changes to show" - **FastHub Version: 4.3.0** **Android Version: 7.0 (SDK: 24)** **Device Information:** - **Manufacturer:** samsung - **Brand:** samsung - **Model:** Galaxy Tab S2 8.0 --- It seems that you added a threshold in a new version and don't show a diff in PR if file contains many changes. This is a horrible decision that significantly decrease the usability. First of all this threshold seems to be very small as most of changes are not being displayed more. Another point is that PR review feature is affected. You sell a PRO feature to perform a code review via app, however how could it be possible to left a comment if diff isn't displayed? Now I'm able to left comments to a few files in only, while for rest I should use the browser. I'm very disappointed by this bug :pensive: For me it's about the current approach in general - in each version you introduce some new "features", that later create the usability issues (minor or major) :cry: It's a bit annoying as I feel like a beta tester.. Of course, you quickly fix them, but for couple of weeks we should live with them :( Could you please consider adding an advanced configuration for such things, so we could quickly disable/tune new features if there issues with them? For instance, it would be enough to tune the threshold value to work around the current issue, rather than be unable to perform a code review for a few weeks.. Thank you for your time and work on this project!
non_process
annoying and blocking too many changes to show fasthub version android version sdk device information manufacturer samsung brand samsung model galaxy tab it seems that you added a threshold in a new version and don t show a diff in pr if file contains many changes this is a horrible decision that significantly decrease the usability first of all this threshold seems to be very small as most of changes are not being displayed more another point is that pr review feature is affected you sell a pro feature to perform a code review via app however how could it be possible to left a comment if diff isn t displayed now i m able to left comments to a few files in only while for rest i should use the browser i m very disappointed by this bug pensive for me it s about the current approach in general in each version you introduce some new features that later create the usability issues minor or major cry it s a bit annoying as i feel like a beta tester of course you quickly fix them but for couple of weeks we should live with them could you please consider adding an advanced configuration for such things so we could quickly disable tune new features if there issues with them for instance it would be enough to tune the threshold value to work around the current issue rather than be unable to perform a code review for a few weeks thank you for your time and work on this project
0
281,303
30,888,605,522
IssuesEvent
2023-08-04 01:34:02
Nivaskumark/kernel_v4.1.15
https://api.github.com/repos/Nivaskumark/kernel_v4.1.15
reopened
WS-2021-0551 (Critical) detected in linuxlinux-4.6
Mend: dependency security vulnerability
## WS-2021-0551 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.1.15/commit/00db4e8795bcbec692fb60b19160bdd763ad42e3">00db4e8795bcbec692fb60b19160bdd763ad42e3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> nfsd: fix use-after-free due to delegation race This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven. This ID is fixed in Linux Kernel version v5.15.8 by commit 148c816f10fd11df27ca6a9b3238cdd42fa72cd3. For more details please see the references link. <p>Publish Date: 2021-12-19 <p>URL: <a href=https://github.com/gregkh/linux/commit/148c816f10fd11df27ca6a9b3238cdd42fa72cd3>WS-2021-0551</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002600">https://osv.dev/vulnerability/GSD-2021-1002600</a></p> <p>Release Date: 2021-12-19</p> <p>Fix Resolution: v5.15.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0551 (Critical) detected in linuxlinux-4.6 - ## WS-2021-0551 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.1.15/commit/00db4e8795bcbec692fb60b19160bdd763ad42e3">00db4e8795bcbec692fb60b19160bdd763ad42e3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> nfsd: fix use-after-free due to delegation race This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven. This ID is fixed in Linux Kernel version v5.15.8 by commit 148c816f10fd11df27ca6a9b3238cdd42fa72cd3. For more details please see the references link. <p>Publish Date: 2021-12-19 <p>URL: <a href=https://github.com/gregkh/linux/commit/148c816f10fd11df27ca6a9b3238cdd42fa72cd3>WS-2021-0551</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002600">https://osv.dev/vulnerability/GSD-2021-1002600</a></p> <p>Release Date: 2021-12-19</p> <p>Fix Resolution: v5.15.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws critical detected in linuxlinux ws critical severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details nfsd fix use after free due to delegation race this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
16,810
22,060,445,295
IssuesEvent
2022-05-30 17:02:56
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Terminal close immediately after opening only if I use zsh
bug confirmation-pending terminal-process
Issue Type: <b>Bug</b> No typical step required, even with a fresh install and an empty .zshrc file the terminal still close after opening. Even in the developer tools there are no errors. I really don't know what to do... VS Code version: Code 1.61.2 (Universal) (6cba118ac49a1b88332f312a8f67186f7f3c1643, 2021-10-19T15:49:28.381Z) OS version: Darwin arm64 20.6.0 Restricted Mode: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Apple M1 (8 x 24)| |GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|4, 4, 4| |Memory (System)|16.00GB (0.62GB free)| |Process Argv|--crash-reporter-id 658b4308-ce3b-4500-a9ec-e674fc511008| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (16)</summary> Extension|Author (truncated)|Version ---|---|--- bracket-pair-colorizer|Coe|1.0.61 doxdocgen|csc|1.3.2 es7-react-js-snippets|dsz|3.1.1 prettier-vscode|esb|9.0.0 auto-rename-tag|for|0.1.9 vscode-jekyll-snippets|gin|0.9.3 solidity|Jua|0.0.134 kite|kit|0.147.0 python|ms-|2021.10.1365161279 vscode-pylance|ms-|2021.10.3 jupyter|ms-|2021.9.1101343141 jupyter-keymap|ms-|1.0.0 jupyter-renderers|ms-|1.0.3 material-icon-theme|PKi|4.10.0 LiveServer|rit|5.6.1 shopify-liquid|sis|2.3.0 </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383cf:30185419 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 pythontb:30283811 pythonptprofiler:30281270 vshan820:30294714 vstes263cf:30335440 vscoreces:30384385 pythondataviewer:30285071 pythonvsuse255:30340121 vscod805cf:30301675 pythonvspyt200:30340761 binariesv615:30325510 vsccppwtct:30382698 bridge0708:30335490 pygetstartedt3:30385195 dockerwalkthru:30377721 bridge0723:30353136 pythonrunftest32:30373476 pythonf5test824:30373475 javagetstartedt:30391933 pythonvspyt187:30373474 vsqsis200cf:30386380 vsaa593:30376534 vssld246cf:30386378 ``` </details> <!-- generated by issue reporter -->
1.0
Terminal close immediately after opening only if I use zsh - Issue Type: <b>Bug</b> No typical step required, even with a fresh install and an empty .zshrc file the terminal still close after opening. Even in the developer tools there are no errors. I really don't know what to do... VS Code version: Code 1.61.2 (Universal) (6cba118ac49a1b88332f312a8f67186f7f3c1643, 2021-10-19T15:49:28.381Z) OS version: Darwin arm64 20.6.0 Restricted Mode: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Apple M1 (8 x 24)| |GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|4, 4, 4| |Memory (System)|16.00GB (0.62GB free)| |Process Argv|--crash-reporter-id 658b4308-ce3b-4500-a9ec-e674fc511008| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (16)</summary> Extension|Author (truncated)|Version ---|---|--- bracket-pair-colorizer|Coe|1.0.61 doxdocgen|csc|1.3.2 es7-react-js-snippets|dsz|3.1.1 prettier-vscode|esb|9.0.0 auto-rename-tag|for|0.1.9 vscode-jekyll-snippets|gin|0.9.3 solidity|Jua|0.0.134 kite|kit|0.147.0 python|ms-|2021.10.1365161279 vscode-pylance|ms-|2021.10.3 jupyter|ms-|2021.9.1101343141 jupyter-keymap|ms-|1.0.0 jupyter-renderers|ms-|1.0.3 material-icon-theme|PKi|4.10.0 LiveServer|rit|5.6.1 shopify-liquid|sis|2.3.0 </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383cf:30185419 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 pythontb:30283811 pythonptprofiler:30281270 vshan820:30294714 vstes263cf:30335440 vscoreces:30384385 pythondataviewer:30285071 pythonvsuse255:30340121 vscod805cf:30301675 pythonvspyt200:30340761 binariesv615:30325510 vsccppwtct:30382698 bridge0708:30335490 pygetstartedt3:30385195 dockerwalkthru:30377721 bridge0723:30353136 pythonrunftest32:30373476 pythonf5test824:30373475 javagetstartedt:30391933 pythonvspyt187:30373474 vsqsis200cf:30386380 vsaa593:30376534 vssld246cf:30386378 ``` </details> <!-- generated by issue reporter -->
process
terminal close immediately after opening only if i use zsh issue type bug no typical step required even with a fresh install and an empty zshrc file the terminal still close after opening even in the developer tools there are no errors i really don t know what to do vs code version code universal os version darwin restricted mode no system info item value cpus apple x gpu status canvas enabled gpu compositing enabled metal disabled off multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled skia renderer disabled off ok video decode enabled webgl enabled enabled load avg memory system free process argv crash reporter id screen reader no vm extensions extension author truncated version bracket pair colorizer coe doxdocgen csc react js snippets dsz prettier vscode esb auto rename tag for vscode jekyll snippets gin solidity jua kite kit python ms vscode pylance ms jupyter ms jupyter keymap ms jupyter renderers ms material icon theme pki liveserver rit shopify liquid sis a b experiments pythontb pythonptprofiler vscoreces pythondataviewer vsccppwtct dockerwalkthru javagetstartedt
1
17,378
23,200,248,194
IssuesEvent
2022-08-01 20:38:52
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
[FALSE-POSITIVE?]
whitelisting process waiting for Mitch
**Domains or links** Please list any domains and links listed here which you believe are a false positive. **https://www.airbnb.com.au/** **More Information** How did you discover your web site or domain was listed here? 1. Website was hacked 2. Other ..... I see no reason to blacklist a very popular accommodation booking website. **Have you requested removal from other sources?** Please include all relevant links to your existing removals / whitelistings. **Additional context** Add any other context about the problem here. :exclamation: We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
1.0
[FALSE-POSITIVE?] - **Domains or links** Please list any domains and links listed here which you believe are a false positive. **https://www.airbnb.com.au/** **More Information** How did you discover your web site or domain was listed here? 1. Website was hacked 2. Other ..... I see no reason to blacklist a very popular accommodation booking website. **Have you requested removal from other sources?** Please include all relevant links to your existing removals / whitelistings. **Additional context** Add any other context about the problem here. :exclamation: We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
process
domains or links please list any domains and links listed here which you believe are a false positive more information how did you discover your web site or domain was listed here website was hacked other i see no reason to blacklist a very popular accommodation booking website have you requested removal from other sources please include all relevant links to your existing removals whitelistings additional context add any other context about the problem here exclamation we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
1
85,210
24,541,461,734
IssuesEvent
2022-10-12 04:24:48
catboost/catboost
https://api.github.com/repos/catboost/catboost
closed
Error while trying to build
build issues
Problem: When I try to run the command `ya make -r catboost/libs/model_interface` I got the error: ``` Traceback (most recent call last): File "devtools/ya/app/init.py", line 657, in configure_exit_interceptor yield File "devtools/ya/app/init.py", line 78, in helper return action(args, **kwargs) File "devtools/ya/entry/entry.py", line 63, in do_main res = handler.handle(handler, args, prefix=['ya']) File "devtools/ya/core/yarg/handler.py", line 222, in handle return handler.handle(self, args[1:], prefix + [name]) File "devtools/ya/core/yarg/dispatch.py", line 38, in handle return self.command().handle(root_handler, args, prefix) File "devtools/ya/core/yarg/handler.py", line 407, in handle return self._action(params) File "devtools/ya/app/init.py", line 110, in helper return action(ctx.params) File "devtools/ya/build/build_handler.py", line 85, in do_ya_make builder = ya_make.YaMake(params, app_ctx) File "devtools/ya/build/ya_make.py", line 924, in init self.ctx = Context(self.opts, app_ctx=app_ctx, graph=graph, tests=tests, stripped_tests=stripped_tests, configure_errors=configure_errors, make_files=make_files, lite_graph=lite_graph) File "devtools/ya/build/ya_make.py", line 594, in init self.graph, self.tests, self.stripped_tests, self.configure_errors, self.make_files = _build_graph_and_tests(self.opts, app_ctx) File "devtools/ya/build/ya_make.py", line 261, in _build_graph_and_tests graph, tests, stripped_tests, gh, make_files = lg.build_graph_and_tests(opts, check=True, ev_listener=ev_listener, display=display) File "devtools/ya/build/graph.py", line 1705, in build_graph_and_tests return _build_graph_and_tests(opts, check, ev_listener, exit_stack, display) File "devtools/ya/build/graph.py", line 2076, in _build_graph_and_tests real_ymake_bin = tools.tool('ymake') File "devtools/ya/yalibrary/tools/init.py", line 220, in tool return toolchain.find(name, with_params, for_platform, cache=cache) File "devtools/ya/yalibrary/tools/init.py", line 155, in find cur_bottle = _bottle(location['bottle'], for_platform) File "devtools/ya/yalibrary/tools/init.py", line 93, in _bottle return _bottler().get(name, for_platform) File "devtools/ya/yalibrary/tools/init.py", line 82, in get return _Bottle(value['formula'], value.get('executable'), for_platform) File "devtools/ya/yalibrary/tools/init.py", line 40, in init self.fetcher = yalibrary.fetcher.ToolChainFetcher(core.config.tool_root(toolscache_version()), self.formula, for_platform) File "devtools/ya/yalibrary/fetcher/init.py", line 586, in ToolChainFetcher impl = _ToolChainSandboxFetcher(root, formula["sandbox_id"], for_platform) TypeError: string indices must be integers, not str ``` catboost version: latest (07.10.22) Operating System: Windows 10 x64 CPU: Intel Core i3-1125G4
1.0
Error while trying to build - Problem: When I try to run the command `ya make -r catboost/libs/model_interface` I got the error: ``` Traceback (most recent call last): File "devtools/ya/app/init.py", line 657, in configure_exit_interceptor yield File "devtools/ya/app/init.py", line 78, in helper return action(args, **kwargs) File "devtools/ya/entry/entry.py", line 63, in do_main res = handler.handle(handler, args, prefix=['ya']) File "devtools/ya/core/yarg/handler.py", line 222, in handle return handler.handle(self, args[1:], prefix + [name]) File "devtools/ya/core/yarg/dispatch.py", line 38, in handle return self.command().handle(root_handler, args, prefix) File "devtools/ya/core/yarg/handler.py", line 407, in handle return self._action(params) File "devtools/ya/app/init.py", line 110, in helper return action(ctx.params) File "devtools/ya/build/build_handler.py", line 85, in do_ya_make builder = ya_make.YaMake(params, app_ctx) File "devtools/ya/build/ya_make.py", line 924, in init self.ctx = Context(self.opts, app_ctx=app_ctx, graph=graph, tests=tests, stripped_tests=stripped_tests, configure_errors=configure_errors, make_files=make_files, lite_graph=lite_graph) File "devtools/ya/build/ya_make.py", line 594, in init self.graph, self.tests, self.stripped_tests, self.configure_errors, self.make_files = _build_graph_and_tests(self.opts, app_ctx) File "devtools/ya/build/ya_make.py", line 261, in _build_graph_and_tests graph, tests, stripped_tests, gh, make_files = lg.build_graph_and_tests(opts, check=True, ev_listener=ev_listener, display=display) File "devtools/ya/build/graph.py", line 1705, in build_graph_and_tests return _build_graph_and_tests(opts, check, ev_listener, exit_stack, display) File "devtools/ya/build/graph.py", line 2076, in _build_graph_and_tests real_ymake_bin = tools.tool('ymake') File "devtools/ya/yalibrary/tools/init.py", line 220, in tool return toolchain.find(name, with_params, for_platform, cache=cache) File "devtools/ya/yalibrary/tools/init.py", line 155, in find cur_bottle = _bottle(location['bottle'], for_platform) File "devtools/ya/yalibrary/tools/init.py", line 93, in _bottle return _bottler().get(name, for_platform) File "devtools/ya/yalibrary/tools/init.py", line 82, in get return _Bottle(value['formula'], value.get('executable'), for_platform) File "devtools/ya/yalibrary/tools/init.py", line 40, in init self.fetcher = yalibrary.fetcher.ToolChainFetcher(core.config.tool_root(toolscache_version()), self.formula, for_platform) File "devtools/ya/yalibrary/fetcher/init.py", line 586, in ToolChainFetcher impl = _ToolChainSandboxFetcher(root, formula["sandbox_id"], for_platform) TypeError: string indices must be integers, not str ``` catboost version: latest (07.10.22) Operating System: Windows 10 x64 CPU: Intel Core i3-1125G4
non_process
error while trying to build problem when i try to run the command ya make r catboost libs model interface i got the error traceback most recent call last file devtools ya app init py line in configure exit interceptor yield file devtools ya app init py line in helper return action args kwargs file devtools ya entry entry py line in do main res handler handle handler args prefix file devtools ya core yarg handler py line in handle return handler handle self args prefix file devtools ya core yarg dispatch py line in handle return self command handle root handler args prefix file devtools ya core yarg handler py line in handle return self action params file devtools ya app init py line in helper return action ctx params file devtools ya build build handler py line in do ya make builder ya make yamake params app ctx file devtools ya build ya make py line in init self ctx context self opts app ctx app ctx graph graph tests tests stripped tests stripped tests configure errors configure errors make files make files lite graph lite graph file devtools ya build ya make py line in init self graph self tests self stripped tests self configure errors self make files build graph and tests self opts app ctx file devtools ya build ya make py line in build graph and tests graph tests stripped tests gh make files lg build graph and tests opts check true ev listener ev listener display display file devtools ya build graph py line in build graph and tests return build graph and tests opts check ev listener exit stack display file devtools ya build graph py line in build graph and tests real ymake bin tools tool ymake file devtools ya yalibrary tools init py line in tool return toolchain find name with params for platform cache cache file devtools ya yalibrary tools init py line in find cur bottle bottle location for platform file devtools ya yalibrary tools init py line in bottle return bottler get name for platform file devtools ya yalibrary tools init py line in get return bottle value value get executable for platform file devtools ya yalibrary tools init py line in init self fetcher yalibrary fetcher toolchainfetcher core config tool root toolscache version self formula for platform file devtools ya yalibrary fetcher init py line in toolchainfetcher impl toolchainsandboxfetcher root formula for platform typeerror string indices must be integers not str catboost version latest operating system windows cpu intel core
0
453,598
13,085,208,943
IssuesEvent
2020-08-02 00:48:13
SpencerTSterling/RatHutProductManager
https://api.github.com/repos/SpencerTSterling/RatHutProductManager
closed
Create a Database class
enhancement high priority
- Add a class to the project ex. ProductDb - Code method stubds ex. Add(), Update(), Delete() ``` csharp throw new NotImplementedException(); ```
1.0
Create a Database class - - Add a class to the project ex. ProductDb - Code method stubds ex. Add(), Update(), Delete() ``` csharp throw new NotImplementedException(); ```
non_process
create a database class add a class to the project ex productdb code method stubds ex add update delete csharp throw new notimplementedexception
0
3,180
6,256,013,537
IssuesEvent
2017-07-14 08:59:03
rogerthat-platform/rogerthat-android-client
https://api.github.com/repos/rogerthat-platform/rogerthat-android-client
closed
IOException in startScanningForQRCodes
priority_major process_wontfix type_bug
``` 2017-07-03 07:56:22 UTC (16 min ago) user008a407f635446b5a4e78d282371d3da@***:osa-loyalty Count: 38 - Version: 0.2732 ``` ``` description: Bug! Unexpected IOException initializing TextureView camera errorMessage: java.io.IOException: setPreviewTexture failed at android.hardware.Camera.setPreviewTexture(Native Method) at com.google.zxing.client.android.camera.QRCodeScannerTextureViewCameraManager.openDriver(QRCodeScannerTextureViewCameraManager.java:60) at com.mobicage.rogerthat.plugins.friends.QRCodeScannerTextureView.startScanningForQRCodes(QRCodeScannerTextureView.java:80) at com.mobicage.rogerthat.plugins.friends.QRCodeScanner.onResume(QRCodeScanner.java:132) at com.mobicage.rogerthat.plugins.friends.ActionScreenActivity.onResume(ActionScreenActivity.java:900) at android.app.Instrumentation.callActivityOnResume(Instrumentation.java:1285) at android.app.Activity.performResume(Activity.java:6730) at android.app.ActivityThread.performResumeActivity(ActivityThread.java:4025) at android.app.ActivityThread.handleResumeActivity(ActivityThread.java:4105) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1779) at android.os.Handler.dispatchMessage(Handler.java:102) at android.os.Looper.loop(Looper.java:145) at android.app.ActivityThread.main(ActivityThread.java:6939) at java.lang.reflect.Method.invoke(Native Method) at java.lang.reflect.Method.invoke(Method.java:372) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1404) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1199) mobicageVersion: 0.2732 occurenceCount: 38 platform: 1 platformVersion: samsung/gtexswifixx/gtexswifi:5.1.1/LMY47V/T280XXU0AQA4:user/release-keys (-) 22 (-) SM-T280 ```
1.0
IOException in startScanningForQRCodes - ``` 2017-07-03 07:56:22 UTC (16 min ago) user008a407f635446b5a4e78d282371d3da@***:osa-loyalty Count: 38 - Version: 0.2732 ``` ``` description: Bug! Unexpected IOException initializing TextureView camera errorMessage: java.io.IOException: setPreviewTexture failed at android.hardware.Camera.setPreviewTexture(Native Method) at com.google.zxing.client.android.camera.QRCodeScannerTextureViewCameraManager.openDriver(QRCodeScannerTextureViewCameraManager.java:60) at com.mobicage.rogerthat.plugins.friends.QRCodeScannerTextureView.startScanningForQRCodes(QRCodeScannerTextureView.java:80) at com.mobicage.rogerthat.plugins.friends.QRCodeScanner.onResume(QRCodeScanner.java:132) at com.mobicage.rogerthat.plugins.friends.ActionScreenActivity.onResume(ActionScreenActivity.java:900) at android.app.Instrumentation.callActivityOnResume(Instrumentation.java:1285) at android.app.Activity.performResume(Activity.java:6730) at android.app.ActivityThread.performResumeActivity(ActivityThread.java:4025) at android.app.ActivityThread.handleResumeActivity(ActivityThread.java:4105) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1779) at android.os.Handler.dispatchMessage(Handler.java:102) at android.os.Looper.loop(Looper.java:145) at android.app.ActivityThread.main(ActivityThread.java:6939) at java.lang.reflect.Method.invoke(Native Method) at java.lang.reflect.Method.invoke(Method.java:372) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1404) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1199) mobicageVersion: 0.2732 occurenceCount: 38 platform: 1 platformVersion: samsung/gtexswifixx/gtexswifi:5.1.1/LMY47V/T280XXU0AQA4:user/release-keys (-) 22 (-) SM-T280 ```
process
ioexception in startscanningforqrcodes utc min ago osa loyalty count version description bug unexpected ioexception initializing textureview camera errormessage java io ioexception setpreviewtexture failed at android hardware camera setpreviewtexture native method at com google zxing client android camera qrcodescannertextureviewcameramanager opendriver qrcodescannertextureviewcameramanager java at com mobicage rogerthat plugins friends qrcodescannertextureview startscanningforqrcodes qrcodescannertextureview java at com mobicage rogerthat plugins friends qrcodescanner onresume qrcodescanner java at com mobicage rogerthat plugins friends actionscreenactivity onresume actionscreenactivity java at android app instrumentation callactivityonresume instrumentation java at android app activity performresume activity java at android app activitythread performresumeactivity activitythread java at android app activitythread handleresumeactivity activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java mobicageversion occurencecount platform platformversion samsung gtexswifixx gtexswifi user release keys sm
1
4,362
7,260,514,351
IssuesEvent
2018-02-18 10:53:27
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] New algorithm to extract specific nodes
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/82312e10e472fb91a5b5c0253456130dea86521e by nyalldawson This algorithm allows you to extract specific nodes from geometries. Eg you can extract the first or last node in the geometry. The algorithm accepts a comma separated list of node indices to extract, eg 0 = first node, 1 = second node, etc. Negative indices can be used to extract nodes from the end of the geometry. Eg -1 = last node, -2 = second last node.
1.0
[FEATURE][processing] New algorithm to extract specific nodes - Original commit: https://github.com/qgis/QGIS/commit/82312e10e472fb91a5b5c0253456130dea86521e by nyalldawson This algorithm allows you to extract specific nodes from geometries. Eg you can extract the first or last node in the geometry. The algorithm accepts a comma separated list of node indices to extract, eg 0 = first node, 1 = second node, etc. Negative indices can be used to extract nodes from the end of the geometry. Eg -1 = last node, -2 = second last node.
process
new algorithm to extract specific nodes original commit by nyalldawson this algorithm allows you to extract specific nodes from geometries eg you can extract the first or last node in the geometry the algorithm accepts a comma separated list of node indices to extract eg first node second node etc negative indices can be used to extract nodes from the end of the geometry eg last node second last node
1
7,571
10,684,626,204
IssuesEvent
2019-10-22 10:54:13
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
Proposed obsoletion multi-organism transport
multi-species process obsoletion
Dear all, The proposal has been made to obsolete the following terms: GO:1902537 multi-organism pinocytosis GO:1990967 multi-organism toxin transport GO:1902586 multi-organism intercellular transport GO:1902588 multi-organism plasmodesmata-mediated intercellular transport The reason for the obsoletion is that only host proteins take part in these processes, so it is NOT a multi-organism process. There are no annotations, no mappings to those terms. Those terms are not present in any slims. Any comments can be added to the issue: https://github.com/geneontology/go-ontology/issues We are opening a comment period for this proposed obsoletion. We’d like to proceed and obsolete this term on October 28th, 2019. Unless objections are received by October 28th 2019, we will assume that you agree to this change. Thanks, Pascale
1.0
Proposed obsoletion multi-organism transport - Dear all, The proposal has been made to obsolete the following terms: GO:1902537 multi-organism pinocytosis GO:1990967 multi-organism toxin transport GO:1902586 multi-organism intercellular transport GO:1902588 multi-organism plasmodesmata-mediated intercellular transport The reason for the obsoletion is that only host proteins take part in these processes, so it is NOT a multi-organism process. There are no annotations, no mappings to those terms. Those terms are not present in any slims. Any comments can be added to the issue: https://github.com/geneontology/go-ontology/issues We are opening a comment period for this proposed obsoletion. We’d like to proceed and obsolete this term on October 28th, 2019. Unless objections are received by October 28th 2019, we will assume that you agree to this change. Thanks, Pascale
process
proposed obsoletion multi organism transport dear all the proposal has been made to obsolete the following terms go multi organism pinocytosis go multi organism toxin transport go multi organism intercellular transport go multi organism plasmodesmata mediated intercellular transport the reason for the obsoletion is that only host proteins take part in these processes so it is not a multi organism process there are no annotations no mappings to those terms those terms are not present in any slims any comments can be added to the issue we are opening a comment period for this proposed obsoletion we’d like to proceed and obsolete this term on october unless objections are received by october we will assume that you agree to this change thanks pascale
1
70,019
30,531,613,739
IssuesEvent
2023-07-19 14:36:28
Ocelot-Social-Community/Ocelot-Social
https://api.github.com/repos/Ocelot-Social-Community/Ocelot-Social
closed
🚀 [Feature][Webapp] Chat room overview: last message text should be a line with ~50chars
feature service: webapp features_chat chat
<!-- You can find the latest issue templates here https://github.com/ulfgebhardt/issue-templates --> ## 🚀 Feature <!-- Give a short summary of the Feature. Use Screenshots if you want. --> ![image](https://github.com/Ocelot-Social-Community/Ocelot-Social/assets/33051975/3e56bdc4-765c-4fc4-bd90-69f96dc077fe) ## 🤖 ToDo - [ ] Compute the last message teaser - [ ] Update chat room last message slot
1.0
🚀 [Feature][Webapp] Chat room overview: last message text should be a line with ~50chars - <!-- You can find the latest issue templates here https://github.com/ulfgebhardt/issue-templates --> ## 🚀 Feature <!-- Give a short summary of the Feature. Use Screenshots if you want. --> ![image](https://github.com/Ocelot-Social-Community/Ocelot-Social/assets/33051975/3e56bdc4-765c-4fc4-bd90-69f96dc077fe) ## 🤖 ToDo - [ ] Compute the last message teaser - [ ] Update chat room last message slot
non_process
🚀 chat room overview last message text should be a line with 🚀 feature 🤖 todo compute the last message teaser update chat room last message slot
0
15,390
19,571,819,885
IssuesEvent
2022-01-04 10:51:13
Kernem/FeRSS-Core
https://api.github.com/repos/Kernem/FeRSS-Core
closed
Filter RSS content by length
post-processing filter
RSS content should be filterable by length, ideally fetching and processing should stop processing some content if it falls outside of the length
1.0
Filter RSS content by length - RSS content should be filterable by length, ideally fetching and processing should stop processing some content if it falls outside of the length
process
filter rss content by length rss content should be filterable by length ideally fetching and processing should stop processing some content if it falls outside of the length
1
495,623
14,285,286,249
IssuesEvent
2020-11-23 13:43:04
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.dailymail.co.uk - video or audio doesn't play
browser-firefox engine-gecko ml-needsdiagnosis-false priority-important
<!-- @browser: Firefox 84.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/62309 --> **URL**: https://www.dailymail.co.uk/ushome/index.html **Browser / Version**: Firefox 84.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Video appears, but will not permit it to play. Also, for some news stories, at times there is no accompanying photograph on the main page. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.dailymail.co.uk - video or audio doesn't play - <!-- @browser: Firefox 84.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/62309 --> **URL**: https://www.dailymail.co.uk/ushome/index.html **Browser / Version**: Firefox 84.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Video appears, but will not permit it to play. Also, for some news stories, at times there is no accompanying photograph on the main page. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
video or audio doesn t play url browser version firefox operating system windows tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce video appears but will not permit it to play also for some news stories at times there is no accompanying photograph on the main page browser configuration none from with ❤️
0
4,563
7,393,696,829
IssuesEvent
2018-03-17 00:47:07
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Possible to import XML doc of Azure agent IPs?
app-service-web cxp in-process product-question triaged
We need to white-list the Azure IPs used for running performance tests against our web app. We cannot find a definitive answer on which IPs these are. So, my though was to import the Azure IP data ranges XML. Thoughts? https://www.microsoft.com/en-us/download/confirmation.aspx?id=41653 Joey --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d90bbf62-44f9-30a6-d6ea-bbf776f6efc5 * Version Independent ID: b3961eac-6c70-b345-f48c-e4c0ec67059c * Content: [Azure App Service IP Restrictions](https://docs.microsoft.com/en-us/azure/app-service/app-service-ip-restrictions) * Content Source: [articles/app-service/app-service-ip-restrictions.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-ip-restrictions.md) * Service: **app-service-web** * GitHub Login: @btardif * Microsoft Alias: **byvinyal**
1.0
Possible to import XML doc of Azure agent IPs? - We need to white-list the Azure IPs used for running performance tests against our web app. We cannot find a definitive answer on which IPs these are. So, my though was to import the Azure IP data ranges XML. Thoughts? https://www.microsoft.com/en-us/download/confirmation.aspx?id=41653 Joey --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d90bbf62-44f9-30a6-d6ea-bbf776f6efc5 * Version Independent ID: b3961eac-6c70-b345-f48c-e4c0ec67059c * Content: [Azure App Service IP Restrictions](https://docs.microsoft.com/en-us/azure/app-service/app-service-ip-restrictions) * Content Source: [articles/app-service/app-service-ip-restrictions.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-ip-restrictions.md) * Service: **app-service-web** * GitHub Login: @btardif * Microsoft Alias: **byvinyal**
process
possible to import xml doc of azure agent ips we need to white list the azure ips used for running performance tests against our web app we cannot find a definitive answer on which ips these are so my though was to import the azure ip data ranges xml thoughts joey document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service web github login btardif microsoft alias byvinyal
1
9,698
12,700,216,845
IssuesEvent
2020-06-22 15:59:00
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Diagnostic.Process.Start(...) does not Support mailto: protocoll
area-System.Diagnostics.Process untriaged
Porting a .Net Framework App to .Net Core 3.0, I was not able to invoke the default mail client with the `mailto:` protocoll using `System.Diagnostics.Process.Start(...)`. .Net Core `3.0.100-preview5-011568` Windows 10 1809 Outlook 2016 Sampe code to reproduce ```c# var body = "This is a body of a message"; var recipients = String.Join(",", "to@user.com", "alsoTo@user.com"); string mailto = $"mailto:{recipients}?Subject={"Subject of message"}&Body={body}"; mailto = Uri.EscapeUriString(mailto); System.Diagnostics.Process.Start(mailto); ``` The Exception that occures: `System.ComponentModel.Win32Exception: 'Das System kann die angegebene Datei nicht finden.'` (Translated `System can not find the file` not sure if that is the actuall english error message)
1.0
System.Diagnostic.Process.Start(...) does not Support mailto: protocoll - Porting a .Net Framework App to .Net Core 3.0, I was not able to invoke the default mail client with the `mailto:` protocoll using `System.Diagnostics.Process.Start(...)`. .Net Core `3.0.100-preview5-011568` Windows 10 1809 Outlook 2016 Sampe code to reproduce ```c# var body = "This is a body of a message"; var recipients = String.Join(",", "to@user.com", "alsoTo@user.com"); string mailto = $"mailto:{recipients}?Subject={"Subject of message"}&Body={body}"; mailto = Uri.EscapeUriString(mailto); System.Diagnostics.Process.Start(mailto); ``` The Exception that occures: `System.ComponentModel.Win32Exception: 'Das System kann die angegebene Datei nicht finden.'` (Translated `System can not find the file` not sure if that is the actuall english error message)
process
system diagnostic process start does not support mailto protocoll porting a net framework app to net core i was not able to invoke the default mail client with the mailto protocoll using system diagnostics process start net core windows outlook sampe code to reproduce c var body this is a body of a message var recipients string join to user com alsoto user com string mailto mailto recipients subject subject of message body body mailto uri escapeuristring mailto system diagnostics process start mailto the exception that occures system componentmodel das system kann die angegebene datei nicht finden translated system can not find the file not sure if that is the actuall english error message
1
12,813
15,082,920,292
IssuesEvent
2021-02-05 15:10:59
cseelhoff/RimThreaded
https://api.github.com/repos/cseelhoff/RimThreaded
opened
"Insulation (Continued)" frequent errors from doors
1.3.0 - 1.4.0 Bug Confirmed Bug Mod Incompatibility Reproducible
**Describe the bug** IMPORTANT: Please first search existing bugs to ensure you are not creating a duplicate bug report. errors with insulation mod enabled **To Reproduce (VERY IMPORTANT)** Steps to reproduce the behavior: 1. Load a save 2. See errors **Error Log** https://gist.github.com/1163af431c358ee45354c634418805a2 **Mod List** see log https://steamcommunity.com/sharedfiles/filedetails/?id=2196454890&searchtext=insulation **Screenshots** * NA
True
"Insulation (Continued)" frequent errors from doors - **Describe the bug** IMPORTANT: Please first search existing bugs to ensure you are not creating a duplicate bug report. errors with insulation mod enabled **To Reproduce (VERY IMPORTANT)** Steps to reproduce the behavior: 1. Load a save 2. See errors **Error Log** https://gist.github.com/1163af431c358ee45354c634418805a2 **Mod List** see log https://steamcommunity.com/sharedfiles/filedetails/?id=2196454890&searchtext=insulation **Screenshots** * NA
non_process
insulation continued frequent errors from doors describe the bug important please first search existing bugs to ensure you are not creating a duplicate bug report errors with insulation mod enabled to reproduce very important steps to reproduce the behavior load a save see errors error log mod list see log screenshots na
0
22,448
31,168,793,016
IssuesEvent
2023-08-16 22:18:58
NCAR/ucomp-pipeline
https://api.github.com/repos/NCAR/ucomp-pipeline
opened
Reprocess test dates before 0.5.x release/reprocessing
needs testing process
Reprocess the following dates: - [ ] 20210719 - [ ] 20220325 - [ ] 20220523 - [ ] 20220712 Check the following: - no level 2 artifacts - good threshold masking in level 2 products - good field radius masking in level 2 PNGs
1.0
Reprocess test dates before 0.5.x release/reprocessing - Reprocess the following dates: - [ ] 20210719 - [ ] 20220325 - [ ] 20220523 - [ ] 20220712 Check the following: - no level 2 artifacts - good threshold masking in level 2 products - good field radius masking in level 2 PNGs
process
reprocess test dates before x release reprocessing reprocess the following dates check the following no level artifacts good threshold masking in level products good field radius masking in level pngs
1
191,679
14,595,180,811
IssuesEvent
2020-12-20 10:10:17
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
codeamp/circuit: plugins/codeamp/graphql/helpers_test.go; 3 LoC
fresh test tiny
Found a possible issue in [codeamp/circuit](https://www.github.com/codeamp/circuit) at [plugins/codeamp/graphql/helpers_test.go](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to projectEnvironment at line 637 may start a goroutine [Click here to see the code in its original context.](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, projectEnvironment := range projectEnvironments { helper.Resolver.DB.Unscoped().Delete(&projectEnvironment) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 06371150b7aaaf104d41907c6f432a795238bf3c
1.0
codeamp/circuit: plugins/codeamp/graphql/helpers_test.go; 3 LoC - Found a possible issue in [codeamp/circuit](https://www.github.com/codeamp/circuit) at [plugins/codeamp/graphql/helpers_test.go](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to projectEnvironment at line 637 may start a goroutine [Click here to see the code in its original context.](https://github.com/codeamp/circuit/blob/06371150b7aaaf104d41907c6f432a795238bf3c/plugins/codeamp/graphql/helpers_test.go#L636-L638) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, projectEnvironment := range projectEnvironments { helper.Resolver.DB.Unscoped().Delete(&projectEnvironment) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 06371150b7aaaf104d41907c6f432a795238bf3c
non_process
codeamp circuit plugins codeamp graphql helpers test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to projectenvironment at line may start a goroutine click here to show the line s of go which triggered the analyzer go for projectenvironment range projectenvironments helper resolver db unscoped delete projectenvironment leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
36,102
9,745,265,560
IssuesEvent
2019-06-03 09:12:54
groupe-sii/ogham
https://api.github.com/repos/groupe-sii/ogham
opened
Provide configurers for mainly used service providers
API Builder P4 enhancement
Provide a simple automatic way to configure email sending through Gmail, Yahoo, ... It could be something like this with fluent API: ```java .email() .gmail() .account() .username() .applicationPassword() ``` An using properties: ``` ogham.email.gmail.account.username= ogham.email.gmail.account.application-password= ``` And nothing else to do. It will automatically configure correctly the SMTP configuration
1.0
Provide configurers for mainly used service providers - Provide a simple automatic way to configure email sending through Gmail, Yahoo, ... It could be something like this with fluent API: ```java .email() .gmail() .account() .username() .applicationPassword() ``` An using properties: ``` ogham.email.gmail.account.username= ogham.email.gmail.account.application-password= ``` And nothing else to do. It will automatically configure correctly the SMTP configuration
non_process
provide configurers for mainly used service providers provide a simple automatic way to configure email sending through gmail yahoo it could be something like this with fluent api java email gmail account username applicationpassword an using properties ogham email gmail account username ogham email gmail account application password and nothing else to do it will automatically configure correctly the smtp configuration
0
18,205
24,260,500,334
IssuesEvent
2022-09-27 22:06:51
python/cpython
https://api.github.com/repos/python/cpython
closed
concurrent.futures.as_completed() installs waiters for already completed Futures
performance stdlib pending expert-multiprocessing
BPO | [20297](https://bugs.python.org/issue20297) --- | :--- Nosy | @brianquinlan, @mdickinson, @vstinner Files | <li>[as_completed_proposed.py](https://bugs.python.org/file33590/as_completed_proposed.py "Uploaded as text/plain at 2014-01-21.15:19:28 by glangford")</li><li>[as_completed_proposed.patch](https://bugs.python.org/file33654/as_completed_proposed.patch "Uploaded as text/plain at 2014-01-23.13:48:02 by @vstinner")</li><li>[test_dupfuture.py](https://bugs.python.org/file33657/test_dupfuture.py "Uploaded as text/plain at 2014-01-23.14:43:21 by glangford")</li> <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2014-01-18.21:19:35.870> labels = ['library', 'performance'] title = 'concurrent.futures.as_completed() installs waiters for already completed Futures' updated_at = <Date 2014-07-18.16:46:17.165> user = 'https://bugs.python.org/glangford' ``` bugs.python.org fields: ```python activity = <Date 2014-07-18.16:46:17.165> actor = 'glangford' assignee = 'none' closed = False closed_date = None closer = None components = ['Library (Lib)'] creation = <Date 2014-01-18.21:19:35.870> creator = 'glangford' dependencies = [] files = ['33590', '33654', '33657'] hgrepos = [] issue_num = 20297 keywords = ['patch'] message_count = 7.0 messages = ['208418', '208593', '208652', '208921', '208942', '208943', '208974'] nosy_count = 3.0 nosy_names = ['bquinlan', 'mark.dickinson', 'vstinner'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'performance' url = 'https://bugs.python.org/issue20297' versions = ['Python 3.3', 'Python 3.4'] ``` </p></details>
1.0
concurrent.futures.as_completed() installs waiters for already completed Futures - BPO | [20297](https://bugs.python.org/issue20297) --- | :--- Nosy | @brianquinlan, @mdickinson, @vstinner Files | <li>[as_completed_proposed.py](https://bugs.python.org/file33590/as_completed_proposed.py "Uploaded as text/plain at 2014-01-21.15:19:28 by glangford")</li><li>[as_completed_proposed.patch](https://bugs.python.org/file33654/as_completed_proposed.patch "Uploaded as text/plain at 2014-01-23.13:48:02 by @vstinner")</li><li>[test_dupfuture.py](https://bugs.python.org/file33657/test_dupfuture.py "Uploaded as text/plain at 2014-01-23.14:43:21 by glangford")</li> <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2014-01-18.21:19:35.870> labels = ['library', 'performance'] title = 'concurrent.futures.as_completed() installs waiters for already completed Futures' updated_at = <Date 2014-07-18.16:46:17.165> user = 'https://bugs.python.org/glangford' ``` bugs.python.org fields: ```python activity = <Date 2014-07-18.16:46:17.165> actor = 'glangford' assignee = 'none' closed = False closed_date = None closer = None components = ['Library (Lib)'] creation = <Date 2014-01-18.21:19:35.870> creator = 'glangford' dependencies = [] files = ['33590', '33654', '33657'] hgrepos = [] issue_num = 20297 keywords = ['patch'] message_count = 7.0 messages = ['208418', '208593', '208652', '208921', '208942', '208943', '208974'] nosy_count = 3.0 nosy_names = ['bquinlan', 'mark.dickinson', 'vstinner'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'performance' url = 'https://bugs.python.org/issue20297' versions = ['Python 3.3', 'Python 3.4'] ``` </p></details>
process
concurrent futures as completed installs waiters for already completed futures bpo nosy brianquinlan mdickinson vstinner files uploaded as text plain at by glangford uploaded as text plain at by vstinner uploaded as text plain at by glangford note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title concurrent futures as completed installs waiters for already completed futures updated at user bugs python org fields python activity actor glangford assignee none closed false closed date none closer none components creation creator glangford dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type performance url versions
1
432,873
30,297,256,998
IssuesEvent
2023-07-10 00:41:59
RE-M4/PV-Final-2023
https://api.github.com/repos/RE-M4/PV-Final-2023
closed
Agregar documentación al controlador Testimonio
documentation
-Se debe agregar documentación a los métodos del controlador y eliminar líneas de código innecesarias de haberlas.
1.0
Agregar documentación al controlador Testimonio - -Se debe agregar documentación a los métodos del controlador y eliminar líneas de código innecesarias de haberlas.
non_process
agregar documentación al controlador testimonio se debe agregar documentación a los métodos del controlador y eliminar líneas de código innecesarias de haberlas
0
12,984
15,356,566,942
IssuesEvent
2021-03-01 12:36:17
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
closed
Asset: Quickstart tests sometimes timing out
api: cloudasset priority: p1 samples type: process
Sample [CI output](https://source.cloud.google.com/results/invocations/8273bb2e-f90b-4617-92b3-6aaf17a0279c/targets/github%2Fdotnet-docs-samples%2Fasset%2Fquickstart%2FExportAssetsTest/tests) This might just be a matter of increasing timeouts.
1.0
Asset: Quickstart tests sometimes timing out - Sample [CI output](https://source.cloud.google.com/results/invocations/8273bb2e-f90b-4617-92b3-6aaf17a0279c/targets/github%2Fdotnet-docs-samples%2Fasset%2Fquickstart%2FExportAssetsTest/tests) This might just be a matter of increasing timeouts.
process
asset quickstart tests sometimes timing out sample this might just be a matter of increasing timeouts
1
10,678
13,462,471,860
IssuesEvent
2020-09-09 16:07:06
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Add to generation scripts for BCO
Format - RDF Process - dismissed priority
BCO always imports from Darwin Core. Create the dwc.owl file from the normative rdf using the scripts.
1.0
Add to generation scripts for BCO - BCO always imports from Darwin Core. Create the dwc.owl file from the normative rdf using the scripts.
process
add to generation scripts for bco bco always imports from darwin core create the dwc owl file from the normative rdf using the scripts
1
14,900
2,610,838,157
IssuesEvent
2015-02-26 22:46:17
tsgrp/ActiveWizard
https://api.github.com/repos/tsgrp/ActiveWizard
closed
Workflow rules are not evaluating properly
High Priority HTML5 issue
See the screencam below. It looks like the workflow summary page is not displaying the correct users for the Initial QA role. ![wf-rules-eval](https://cloud.githubusercontent.com/assets/1377228/6380123/f1af9ba0-bcfc-11e4-9b38-7cf38de3d993.gif) I see two issues: - [ ] The workflow summary page is displaying the wrong user for the IQA role - [ ] The reason for the approver being selected is wrong (see below) In the screenshot below, it looks like the high level rule is being used to determine the approver. The IQA role should trip with: - High level rule - Priority of Change is not Low - User level rule - Priority of Change is Medium - User is Max ![image](https://cloud.githubusercontent.com/assets/1377228/6380164/3ad20250-bcfd-11e4-89ac-c5e2d01c39a6.png) @mikeblum @dgrumieaux - Not sure if this is a problem for our current client's form, but we'll definitely need this for the HPI 2.3 release since we primarily demo with the Simple CR form.
1.0
Workflow rules are not evaluating properly - See the screencam below. It looks like the workflow summary page is not displaying the correct users for the Initial QA role. ![wf-rules-eval](https://cloud.githubusercontent.com/assets/1377228/6380123/f1af9ba0-bcfc-11e4-9b38-7cf38de3d993.gif) I see two issues: - [ ] The workflow summary page is displaying the wrong user for the IQA role - [ ] The reason for the approver being selected is wrong (see below) In the screenshot below, it looks like the high level rule is being used to determine the approver. The IQA role should trip with: - High level rule - Priority of Change is not Low - User level rule - Priority of Change is Medium - User is Max ![image](https://cloud.githubusercontent.com/assets/1377228/6380164/3ad20250-bcfd-11e4-89ac-c5e2d01c39a6.png) @mikeblum @dgrumieaux - Not sure if this is a problem for our current client's form, but we'll definitely need this for the HPI 2.3 release since we primarily demo with the Simple CR form.
non_process
workflow rules are not evaluating properly see the screencam below it looks like the workflow summary page is not displaying the correct users for the initial qa role i see two issues the workflow summary page is displaying the wrong user for the iqa role the reason for the approver being selected is wrong see below in the screenshot below it looks like the high level rule is being used to determine the approver the iqa role should trip with high level rule priority of change is not low user level rule priority of change is medium user is max mikeblum dgrumieaux not sure if this is a problem for our current client s form but we ll definitely need this for the hpi release since we primarily demo with the simple cr form
0
741,042
25,777,767,815
IssuesEvent
2022-12-09 13:27:06
bounswe/bounswe2022group4
https://api.github.com/repos/bounswe/bounswe2022group4
closed
Frontend: Update Create Comment UI
Category - To Do Priority - High Status: In Progress whom: individual Difficulty - Hard Language - React.js Team - Frontend
I have implemented a create comment structure that only allow user to provide body. Since backend has changed now i need to implement a structure that allow user to provide upvote a comment, downvote a comment and delete a comment. Steps: 1) Make research on material ui modal structure 2) Implementing a new UI for the Create Comment Component 3) Add new fields with functionalities in Create Comment Component Reviewer: @BeratDamar Deadline: 04.12.2022 23.59
1.0
Frontend: Update Create Comment UI - I have implemented a create comment structure that only allow user to provide body. Since backend has changed now i need to implement a structure that allow user to provide upvote a comment, downvote a comment and delete a comment. Steps: 1) Make research on material ui modal structure 2) Implementing a new UI for the Create Comment Component 3) Add new fields with functionalities in Create Comment Component Reviewer: @BeratDamar Deadline: 04.12.2022 23.59
non_process
frontend update create comment ui i have implemented a create comment structure that only allow user to provide body since backend has changed now i need to implement a structure that allow user to provide upvote a comment downvote a comment and delete a comment steps make research on material ui modal structure implementing a new ui for the create comment component add new fields with functionalities in create comment component reviewer beratdamar deadline
0
1,678
4,314,001,904
IssuesEvent
2016-07-22 12:59:34
pelias/api
https://api.github.com/repos/pelias/api
closed
Move back to upstream cluster2 package if our PR is merged
processed
Update package.json to remove the `git://` url used as an interim solution until the upstream PR is closed, see commit: https://github.com/pelias/api/commit/e3efeb66d18374524f89696754db0a024c27dc73 The version number targeted for `cluster2` MUST include this PR: https://github.com/cubejs/cluster2/pull/86
1.0
Move back to upstream cluster2 package if our PR is merged - Update package.json to remove the `git://` url used as an interim solution until the upstream PR is closed, see commit: https://github.com/pelias/api/commit/e3efeb66d18374524f89696754db0a024c27dc73 The version number targeted for `cluster2` MUST include this PR: https://github.com/cubejs/cluster2/pull/86
process
move back to upstream package if our pr is merged update package json to remove the git url used as an interim solution until the upstream pr is closed see commit the version number targeted for must include this pr
1
22,001
30,504,259,779
IssuesEvent
2023-07-18 15:45:39
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
init-npm-por-lean-parte_tres 1.0.1 has 5 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/import-npm/node_modules/init-npm/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/init-npm-por-lean-parte_tres/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/import-npm/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":"\t\tspawn(process.execPath, [path.join(__dirname, 'check.js'), JSON.stringify(this.options)], {\n\t\t\tdetached: true,\n\t\t\tstdio: 'ignore'\n\t\t}).unref();","location":"package/import-npm/node_modules/update-notifier/index.js:97","message":"This package is silently executing another executable"}]}```
1.0
init-npm-por-lean-parte_tres 1.0.1 has 5 guarddog issues - ```{"npm-install-script":[{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/import-npm/node_modules/init-npm/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install \u0026\u0026 npm run prepare:hooks\",","location":"package/import-npm/node_modules/init-npm-por-lean-parte_tres/import-npm/node_modules/axios/package.json:53","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/import-npm/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":"\t\tspawn(process.execPath, [path.join(__dirname, 'check.js'), JSON.stringify(this.options)], {\n\t\t\tdetached: true,\n\t\t\tstdio: 'ignore'\n\t\t}).unref();","location":"package/import-npm/node_modules/update-notifier/index.js:97","message":"This package is silently executing another executable"}]}```
process
init npm por lean parte tres has guarddog issues npm install script npm silent process execution n t t tdetached true n t t tstdio ignore n t t unref location package import npm node modules update notifier index js message this package is silently executing another executable
1
11,413
30,453,357,501
IssuesEvent
2023-07-16 15:21:42
SuperCowPowers/sageworks
https://api.github.com/repos/SuperCowPowers/sageworks
opened
Have a 'refresh now' option for the AWS Service Broker
api aws_service_broker application architecture
When we 'delete' an artifacts on the AWS Dashboard, there's this weird row deletes, row shows back up and finally row goes away. This is because of the 'eventually consistence' nature of the AWS Service Broker. Anyway lets have the view expose the 'force_refresh' option in some way and have the app call it when a delete happens.
1.0
Have a 'refresh now' option for the AWS Service Broker - When we 'delete' an artifacts on the AWS Dashboard, there's this weird row deletes, row shows back up and finally row goes away. This is because of the 'eventually consistence' nature of the AWS Service Broker. Anyway lets have the view expose the 'force_refresh' option in some way and have the app call it when a delete happens.
non_process
have a refresh now option for the aws service broker when we delete an artifacts on the aws dashboard there s this weird row deletes row shows back up and finally row goes away this is because of the eventually consistence nature of the aws service broker anyway lets have the view expose the force refresh option in some way and have the app call it when a delete happens
0
107,075
11,516,690,143
IssuesEvent
2020-02-14 06:04:38
postmanlabs/postman-app-support
https://api.github.com/repos/postmanlabs/postman-app-support
closed
Confusing UX in the `New > Create documentation` flow in the app
Microcopy bug product/documentation
1. Start with a new Postman installation (signed-in to a free account). 2. Create a collection (with name and description) 3. New > Documentation > (Use collection from this workspace) Problem 1: The text still says 'Enter the requests you want to document'. This is relevant for the `Create a new API` tab, not the `Use collection from this workspace` tab. 4. Select the collection you created Problem 2: The name isn't editable, but the description is. Why the discrepancy? The text on the right says 'Enter a title to describe your requests', but I'm not allowed to change the name. Seen on Postman v6.6, Mac native app
1.0
Confusing UX in the `New > Create documentation` flow in the app - 1. Start with a new Postman installation (signed-in to a free account). 2. Create a collection (with name and description) 3. New > Documentation > (Use collection from this workspace) Problem 1: The text still says 'Enter the requests you want to document'. This is relevant for the `Create a new API` tab, not the `Use collection from this workspace` tab. 4. Select the collection you created Problem 2: The name isn't editable, but the description is. Why the discrepancy? The text on the right says 'Enter a title to describe your requests', but I'm not allowed to change the name. Seen on Postman v6.6, Mac native app
non_process
confusing ux in the new create documentation flow in the app start with a new postman installation signed in to a free account create a collection with name and description new documentation use collection from this workspace problem the text still says enter the requests you want to document this is relevant for the create a new api tab not the use collection from this workspace tab select the collection you created problem the name isn t editable but the description is why the discrepancy the text on the right says enter a title to describe your requests but i m not allowed to change the name seen on postman mac native app
0
12,082
14,740,041,711
IssuesEvent
2021-01-07 08:24:46
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Chattanooga - SA Billing - Late Fee Account List
anc-process anp-important ant-bug has attachment
In GitLab by @kdjstudios on Oct 3, 2018, 10:58 [Chattanooga.xlsx](/uploads/f892636fe4033036654065e7216841e7/Chattanooga.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-88683/conversation
1.0
Chattanooga - SA Billing - Late Fee Account List - In GitLab by @kdjstudios on Oct 3, 2018, 10:58 [Chattanooga.xlsx](/uploads/f892636fe4033036654065e7216841e7/Chattanooga.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-88683/conversation
process
chattanooga sa billing late fee account list in gitlab by kdjstudios on oct uploads chattanooga xlsx hd
1
425,496
29,482,405,044
IssuesEvent
2023-06-02 07:05:05
aditya-grover/climate-learn
https://api.github.com/repos/aditya-grover/climate-learn
opened
Statistical Downscaling of other ERA5 Variables
documentation
So I am attempting to downscale ERA5 Sea Surface Temperature Variable, I was following along your tutorial at NeurIPS2022 CCAI. I noticed there you used 5 degree and 2 Degree Resolutions for 2m_temperature, Why is this done? It is not very clear. For Sea Surface Temperature i have data at 0.25 Degree resolution, but do i need a Coarser resolution to get this code to work for my chosen variable?
1.0
Statistical Downscaling of other ERA5 Variables - So I am attempting to downscale ERA5 Sea Surface Temperature Variable, I was following along your tutorial at NeurIPS2022 CCAI. I noticed there you used 5 degree and 2 Degree Resolutions for 2m_temperature, Why is this done? It is not very clear. For Sea Surface Temperature i have data at 0.25 Degree resolution, but do i need a Coarser resolution to get this code to work for my chosen variable?
non_process
statistical downscaling of other variables so i am attempting to downscale sea surface temperature variable i was following along your tutorial at ccai i noticed there you used degree and degree resolutions for temperature why is this done it is not very clear for sea surface temperature i have data at degree resolution but do i need a coarser resolution to get this code to work for my chosen variable
0
388,841
26,784,136,055
IssuesEvent
2023-02-01 00:27:35
automatic-ripping-machine/automatic-ripping-machine
https://api.github.com/repos/automatic-ripping-machine/automatic-ripping-machine
closed
No NVENC hardware encoding option with 1.3.x handbrake installed with the script on Ubuntu 20.04.x
documentation Resolved
- [ x] I have checked the wiki - [x ] I have searched for similar issues **Describe the bug** So it's no really a bug, just something that I have found. I have a GTX 1060 NVENC compatible card seen by nvidia-smi tool but handbrake does not list the nvenc encoding option, juste like the hardware encoding is not available with the installed version **To Reproduce** Steps to reproduce the behavior: 1. Install an ubuntu 20.04.x vm in promox (pci pasthrought should be enable and the card should be listed when executing the nvida-smi tool) 2. Install ARM using the script 3. use Handbranke-cli to check the encoding options ### Environment Ubuntu 20.04.5 LTS ARM: 2.6.10
1.0
No NVENC hardware encoding option with 1.3.x handbrake installed with the script on Ubuntu 20.04.x - - [ x] I have checked the wiki - [x ] I have searched for similar issues **Describe the bug** So it's no really a bug, just something that I have found. I have a GTX 1060 NVENC compatible card seen by nvidia-smi tool but handbrake does not list the nvenc encoding option, juste like the hardware encoding is not available with the installed version **To Reproduce** Steps to reproduce the behavior: 1. Install an ubuntu 20.04.x vm in promox (pci pasthrought should be enable and the card should be listed when executing the nvida-smi tool) 2. Install ARM using the script 3. use Handbranke-cli to check the encoding options ### Environment Ubuntu 20.04.5 LTS ARM: 2.6.10
non_process
no nvenc hardware encoding option with x handbrake installed with the script on ubuntu x i have checked the wiki i have searched for similar issues describe the bug so it s no really a bug just something that i have found i have a gtx nvenc compatible card seen by nvidia smi tool but handbrake does not list the nvenc encoding option juste like the hardware encoding is not available with the installed version to reproduce steps to reproduce the behavior install an ubuntu x vm in promox pci pasthrought should be enable and the card should be listed when executing the nvida smi tool install arm using the script use handbranke cli to check the encoding options environment ubuntu lts arm
0
16,280
20,904,904,947
IssuesEvent
2022-03-24 00:19:46
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
closed
Update CI for the BladeHawk ruleset release
work-in-progress issue-processing-state-06
The [release](https://github.com/quark-engine/quark-rules/issues/18) of the BladeHawk ruleset introduced more rules into the quark-rule repo. Since [our smoke test](https://github.com/quark-engine/quark-engine/blob/master/.github/workflows/smoke_test.yml) uses this repo for the accuracy checks, this release has made it inaccurate. Thus, we need to adjust it as well.
1.0
Update CI for the BladeHawk ruleset release - The [release](https://github.com/quark-engine/quark-rules/issues/18) of the BladeHawk ruleset introduced more rules into the quark-rule repo. Since [our smoke test](https://github.com/quark-engine/quark-engine/blob/master/.github/workflows/smoke_test.yml) uses this repo for the accuracy checks, this release has made it inaccurate. Thus, we need to adjust it as well.
process
update ci for the bladehawk ruleset release the of the bladehawk ruleset introduced more rules into the quark rule repo since uses this repo for the accuracy checks this release has made it inaccurate thus we need to adjust it as well
1
19,043
25,043,297,496
IssuesEvent
2022-11-05 00:33:16
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
"Instructions for approvers" text box doesn't seem to do anything
product-feedback cba Pri1 azure-devops-pipelines/svc azure-devops-pipelines-process/subsvc
When defining an approval on a secured resource (environment, service connection, _et al_) in Azure DevOps Services, there's a "Instructions to approvers (optional)" text box. This feature doesn't seem to do anything. Text entered in this box is not included in the emails generated by Azure DevOps Services to the approvers, nor is it displayed in the flyout panel on a pipeline run that activates that approval. the box is too small to contain any meaningful amount of text, so it doesn't seem that this intended as a way to document the approval itself. Is this feature supposed to do anything? --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 624da995-5b05-bb05-0afc-c42a1ff5ab87 * Version Independent ID: 3cf8daeb-8d99-845e-dfd8-d0be5f65182c * Content: [Pipeline deployment approvals - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass) * Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/approvals.md) * Service: **azure-devops-pipelines** * Sub-service: **azure-devops-pipelines-process** * GitHub Login: @shashban * Microsoft Alias: **shashban**
1.0
"Instructions for approvers" text box doesn't seem to do anything - When defining an approval on a secured resource (environment, service connection, _et al_) in Azure DevOps Services, there's a "Instructions to approvers (optional)" text box. This feature doesn't seem to do anything. Text entered in this box is not included in the emails generated by Azure DevOps Services to the approvers, nor is it displayed in the flyout panel on a pipeline run that activates that approval. the box is too small to contain any meaningful amount of text, so it doesn't seem that this intended as a way to document the approval itself. Is this feature supposed to do anything? --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 624da995-5b05-bb05-0afc-c42a1ff5ab87 * Version Independent ID: 3cf8daeb-8d99-845e-dfd8-d0be5f65182c * Content: [Pipeline deployment approvals - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass) * Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/approvals.md) * Service: **azure-devops-pipelines** * Sub-service: **azure-devops-pipelines-process** * GitHub Login: @shashban * Microsoft Alias: **shashban**
process
instructions for approvers text box doesn t seem to do anything when defining an approval on a secured resource environment service connection et al in azure devops services there s a instructions to approvers optional text box this feature doesn t seem to do anything text entered in this box is not included in the emails generated by azure devops services to the approvers nor is it displayed in the flyout panel on a pipeline run that activates that approval the box is too small to contain any meaningful amount of text so it doesn t seem that this intended as a way to document the approval itself is this feature supposed to do anything document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service azure devops pipelines sub service azure devops pipelines process github login shashban microsoft alias shashban
1
1,119
3,592,791,528
IssuesEvent
2016-02-01 17:15:31
DoSomething/quasar
https://api.github.com/repos/DoSomething/quasar
closed
Get Answers from Mobile Commons re: Ingestion
#discussion #processing BLOCKER Data 1.0 enhancement small
- Is the "First Seen ID" the opt in path or the campaign ID? - What’s the update on us accessing the database (vs. pinging the API)? - Can Mobile Commons build something/help with manually subscribing the cohort of people who don’t text in the campaign keyword but express interest in the campaign? - Is the current version of API the latest version? Are there any changes coming down the pipeline? Is documentation up to date?(edited) - Can we query the “description” field in a way that is meaningful to us?
1.0
Get Answers from Mobile Commons re: Ingestion - - Is the "First Seen ID" the opt in path or the campaign ID? - What’s the update on us accessing the database (vs. pinging the API)? - Can Mobile Commons build something/help with manually subscribing the cohort of people who don’t text in the campaign keyword but express interest in the campaign? - Is the current version of API the latest version? Are there any changes coming down the pipeline? Is documentation up to date?(edited) - Can we query the “description” field in a way that is meaningful to us?
process
get answers from mobile commons re ingestion is the first seen id the opt in path or the campaign id what’s the update on us accessing the database vs pinging the api can mobile commons build something help with manually subscribing the cohort of people who don’t text in the campaign keyword but express interest in the campaign is the current version of api the latest version are there any changes coming down the pipeline is documentation up to date edited can we query the “description” field in a way that is meaningful to us
1
190,722
15,255,225,880
IssuesEvent
2021-02-20 15:16:03
getpatchwork/patchwork
https://api.github.com/repos/getpatchwork/patchwork
closed
patchwork fails to parse patch with unicode text
bug documentation
The sourceware instance of patchwork (based on the stable 2.2 branch) fails to parse this patch posted on the list: https://sourceware.org/pipermail/libc-alpha/2021-January/121972.html It fails with the following trace: ``` Error when parsing incoming email: OperationalError(1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1") Traceback (most recent call last): File "{redacted}/django/db/backends/utils.py", line 84, in _execute return self.cursor.execute(sql, params) File "{redacted}/django/db/backends/mysql/base.py", line 71, in execute return self.cursor.execute(query, args) File "{redacted}/MySQLdb/cursors.py", line 209, in execute res = self._query(query) File "{redacted}/MySQLdb/cursors.py", line 315, in _query db.query(q) File "{redacted}/MySQLdb/connections.py", line 239, in query _mysql.connection.query(self, query) MySQLdb._exceptions.OperationalError: (1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1") ```
1.0
patchwork fails to parse patch with unicode text - The sourceware instance of patchwork (based on the stable 2.2 branch) fails to parse this patch posted on the list: https://sourceware.org/pipermail/libc-alpha/2021-January/121972.html It fails with the following trace: ``` Error when parsing incoming email: OperationalError(1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1") Traceback (most recent call last): File "{redacted}/django/db/backends/utils.py", line 84, in _execute return self.cursor.execute(sql, params) File "{redacted}/django/db/backends/mysql/base.py", line 71, in execute return self.cursor.execute(query, args) File "{redacted}/MySQLdb/cursors.py", line 209, in execute res = self._query(query) File "{redacted}/MySQLdb/cursors.py", line 315, in _query db.query(q) File "{redacted}/MySQLdb/connections.py", line 239, in query _mysql.connection.query(self, query) MySQLdb._exceptions.OperationalError: (1366, "Incorrect string value: '\\xD0\\x92\\xD1\\x8B\\xD0\\xB4...' for column `patchwork`.`patchwork_patch`.`diff` at row 1") ```
non_process
patchwork fails to parse patch with unicode text the sourceware instance of patchwork based on the stable branch fails to parse this patch posted on the list it fails with the following trace error when parsing incoming email operationalerror incorrect string value for column patchwork patchwork patch diff at row traceback most recent call last file redacted django db backends utils py line in execute return self cursor execute sql params file redacted django db backends mysql base py line in execute return self cursor execute query args file redacted mysqldb cursors py line in execute res self query query file redacted mysqldb cursors py line in query db query q file redacted mysqldb connections py line in query mysql connection query self query mysqldb exceptions operationalerror incorrect string value for column patchwork patchwork patch diff at row
0
225,913
24,911,599,670
IssuesEvent
2022-10-29 23:13:09
Piotr1215/dca-prep-kit
https://api.github.com/repos/Piotr1215/dca-prep-kit
closed
core-2.1.0.tgz: 4 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>core-2.1.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-7753](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-0.0.1.tgz | Transitive | N/A | &#10060; | | [CVE-2022-3517](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | minimatch-3.0.4.tgz | Transitive | N/A | &#10060; | | [CVE-2022-37599](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | loader-utils-2.0.2.tgz | Transitive | N/A | &#10060; | | [CVE-2022-33987](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | got-9.6.0.tgz | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7753</summary> ### Vulnerable Library - <b>trim-0.0.1.tgz</b></p> <p>Trim string whitespace</p> <p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/remark-parse/node_modules/trim/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - mdx-loader-2.1.0.tgz - mdx-1.6.22.tgz - remark-parse-8.0.3.tgz - :x: **trim-0.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim(). <p>Publish Date: 2020-10-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-10-27</p> <p>Fix Resolution: trim - 0.0.3</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-3517</summary> ### Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/serve-handler/node_modules/minimatch/package.json,/node_modules/recursive-readdir/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - react-dev-utils-12.0.1.tgz - recursive-readdir-2.2.2.tgz - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. <p>Publish Date: 2022-10-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-17</p> <p>Fix Resolution: minimatch - 3.0.5</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-37599</summary> ### Vulnerable Library - <b>loader-utils-2.0.2.tgz</b></p> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - file-loader-6.2.0.tgz - :x: **loader-utils-2.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js. <p>Publish Date: 2022-10-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary> ### Vulnerable Library - <b>got-9.6.0.tgz</b></p> <p>Simplified HTTP requests</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-9.6.0.tgz">https://registry.npmjs.org/got/-/got-9.6.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/got/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - update-notifier-5.1.0.tgz - latest-version-5.1.0.tgz - package-json-6.5.0.tgz - :x: **got-9.6.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket. <p>Publish Date: 2022-06-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p> <p>Release Date: 2022-06-18</p> <p>Fix Resolution: got - 11.8.5,12.1.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
core-2.1.0.tgz: 4 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>core-2.1.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-7753](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-0.0.1.tgz | Transitive | N/A | &#10060; | | [CVE-2022-3517](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | minimatch-3.0.4.tgz | Transitive | N/A | &#10060; | | [CVE-2022-37599](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | loader-utils-2.0.2.tgz | Transitive | N/A | &#10060; | | [CVE-2022-33987](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | got-9.6.0.tgz | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7753</summary> ### Vulnerable Library - <b>trim-0.0.1.tgz</b></p> <p>Trim string whitespace</p> <p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/remark-parse/node_modules/trim/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - mdx-loader-2.1.0.tgz - mdx-1.6.22.tgz - remark-parse-8.0.3.tgz - :x: **trim-0.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim(). <p>Publish Date: 2020-10-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-10-27</p> <p>Fix Resolution: trim - 0.0.3</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-3517</summary> ### Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/serve-handler/node_modules/minimatch/package.json,/node_modules/recursive-readdir/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - react-dev-utils-12.0.1.tgz - recursive-readdir-2.2.2.tgz - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. <p>Publish Date: 2022-10-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-17</p> <p>Fix Resolution: minimatch - 3.0.5</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-37599</summary> ### Vulnerable Library - <b>loader-utils-2.0.2.tgz</b></p> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - file-loader-6.2.0.tgz - :x: **loader-utils-2.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js. <p>Publish Date: 2022-10-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary> ### Vulnerable Library - <b>got-9.6.0.tgz</b></p> <p>Simplified HTTP requests</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-9.6.0.tgz">https://registry.npmjs.org/got/-/got-9.6.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/got/package.json</p> <p> Dependency Hierarchy: - core-2.1.0.tgz (Root Library) - update-notifier-5.1.0.tgz - latest-version-5.1.0.tgz - package-json-6.5.0.tgz - :x: **got-9.6.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Piotr1215/dca-prep-kit/commit/35d1dcb73d5933d27815314d8aed54b3976b2791">35d1dcb73d5933d27815314d8aed54b3976b2791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket. <p>Publish Date: 2022-06-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p> <p>Release Date: 2022-06-18</p> <p>Fix Resolution: got - 11.8.5,12.1.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_process
core tgz vulnerabilities highest severity is autoclosed vulnerable library core tgz path to dependency file package json path to vulnerable library node modules loader utils package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high trim tgz transitive n a high minimatch tgz transitive n a high loader utils tgz transitive n a medium got tgz transitive n a details cve vulnerable library trim tgz trim string whitespace library home page a href path to dependency file package json path to vulnerable library node modules remark parse node modules trim package json dependency hierarchy core tgz root library mdx loader tgz mdx tgz remark parse tgz x trim tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package trim are vulnerable to regular expression denial of service redos via trim publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution trim step up your open source security game with mend cve vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules serve handler node modules minimatch package json node modules recursive readdir node modules minimatch package json dependency hierarchy core tgz root library react dev utils tgz recursive readdir tgz x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch step up your open source security game with mend cve vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy core tgz root library file loader tgz x loader utils tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the resourcepath variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend cve vulnerable library got tgz simplified http requests library home page a href path to dependency file package json path to vulnerable library node modules got package json dependency hierarchy core tgz root library update notifier tgz latest version tgz package json tgz x got tgz vulnerable library found in head commit a href found in base branch master vulnerability details the got package before also fixed in for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend
0
6,109
5,285,190,544
IssuesEvent
2017-02-08 03:34:46
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
opened
High memory usage when applying deployment manifest retention policy
area/scale feature/performance
As reported here: http://help.octopusdeploy.com/discussions/problems/51077 Jeff's (the reporting customer) diagnosis seems correct. We are loading _all_ variable sets for all completed and failed deployments. We should stream these.
True
High memory usage when applying deployment manifest retention policy - As reported here: http://help.octopusdeploy.com/discussions/problems/51077 Jeff's (the reporting customer) diagnosis seems correct. We are loading _all_ variable sets for all completed and failed deployments. We should stream these.
non_process
high memory usage when applying deployment manifest retention policy as reported here jeff s the reporting customer diagnosis seems correct we are loading all variable sets for all completed and failed deployments we should stream these
0
89,654
25,867,674,376
IssuesEvent
2022-12-13 22:30:57
USGCRP/gcis
https://api.github.com/repos/USGCRP/gcis
closed
Handle special characters
help wanted type content type technical priority medium on hold for rebuild
Example 1 : When an organization with special character(s) is added as an org title, the automatically generated identifier ignores the special characters. Example 2: Person name with special character(s) is currently hard to find using the search query unless the last name or first name doesn't have special characters in it. We should either teach the system to handle special characters or ignore using them to avoid redundancy.
1.0
Handle special characters - Example 1 : When an organization with special character(s) is added as an org title, the automatically generated identifier ignores the special characters. Example 2: Person name with special character(s) is currently hard to find using the search query unless the last name or first name doesn't have special characters in it. We should either teach the system to handle special characters or ignore using them to avoid redundancy.
non_process
handle special characters example when an organization with special character s is added as an org title the automatically generated identifier ignores the special characters example person name with special character s is currently hard to find using the search query unless the last name or first name doesn t have special characters in it we should either teach the system to handle special characters or ignore using them to avoid redundancy
0
9,397
2,615,147,568
IssuesEvent
2015-03-01 06:23:53
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
Add case studies as new resources on front page
auto-migrated Milestone-3 Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `v...@google.com` on 30 Sep 2010 at 8:33
1.0
Add case studies as new resources on front page - ``` What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `v...@google.com` on 30 Sep 2010 at 8:33
non_process
add case studies as new resources on front page what steps will reproduce the problem what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by v google com on sep at
0
315,209
9,607,783,061
IssuesEvent
2019-05-11 22:16:19
x-klanas/Wrath
https://api.github.com/repos/x-klanas/Wrath
opened
SteamVR actions
2 points high priority user story
As a player I want to be able to customize the controller bindings. - [ ] the required controller bindings must be created and setup appropriately - [ ] the default bindings for as many controllers as possible must be set
1.0
SteamVR actions - As a player I want to be able to customize the controller bindings. - [ ] the required controller bindings must be created and setup appropriately - [ ] the default bindings for as many controllers as possible must be set
non_process
steamvr actions as a player i want to be able to customize the controller bindings the required controller bindings must be created and setup appropriately the default bindings for as many controllers as possible must be set
0
141,201
11,403,655,731
IssuesEvent
2020-01-31 07:50:56
NethermindEth/nethermind
https://api.github.com/repos/NethermindEth/nethermind
closed
Not all ethereum tests running on MacOs and Linux
devops test
![image](https://user-images.githubusercontent.com/498913/73134646-5022e880-4039-11ea-81b5-ecf595ae43b4.png) ![image](https://user-images.githubusercontent.com/498913/73134654-603ac800-4039-11ea-98b3-d53954265bf5.png) I have these failing tests on one of the branches (proofs). It fails on Windows build but passed on Linux and MacOS. After short investigation I can see that Linux and MacOS simply do not run some of the tests.
1.0
Not all ethereum tests running on MacOs and Linux - ![image](https://user-images.githubusercontent.com/498913/73134646-5022e880-4039-11ea-81b5-ecf595ae43b4.png) ![image](https://user-images.githubusercontent.com/498913/73134654-603ac800-4039-11ea-98b3-d53954265bf5.png) I have these failing tests on one of the branches (proofs). It fails on Windows build but passed on Linux and MacOS. After short investigation I can see that Linux and MacOS simply do not run some of the tests.
non_process
not all ethereum tests running on macos and linux i have these failing tests on one of the branches proofs it fails on windows build but passed on linux and macos after short investigation i can see that linux and macos simply do not run some of the tests
0
11,951
14,713,191,838
IssuesEvent
2021-01-05 10:00:47
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
unable to run migrations in an empty PostGIS db
bug/2-confirmed kind/bug process/candidate team/migrations
<!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description Copied from https://github.com/prisma/prisma/discussions/4732 <!-- A clear and concise description of what the bug is. --> I'm struggling to use the migration tool after I upgraded to 2.13.0. I have a few spatial columns that I manage with scripts besides the Prisma migration tool. My pipeline used to be this: 1. apply Prisma migrations 2. run manual migrations and add / alter spatial columns if needed but with 2.13.0, I can no longer run any Prisma migration on a fresh PostGIS database because of this error: ``` Error: P3005 The database schema for `localhost:5432` is not empty. Read more about how to baseline an existing production database: https://pris.ly/d/migrate-baseline ``` Enabling PostGIS creates a `spatial_ref_sys` table, so technically the database is not empty, but it should not affect migrations. I know Prisma does not yet support the PostGIS extension, but I'm looking for a workaround. I'm also a Nexus and nexus-plugin-prisma user which doesn't support 2.12.x, so I can't downgrade reliably. ## How to reproduce Steps to reproduce the behavior: 1. Have an empty PostGIS database (I used https://hub.docker.com/r/postgis/postgis) 1. Run `prisma migrate dev --preview-feature` 1. See error ## Expected behavior Migrations run without an error. ## Environment & setup <!-- In which environment does the problem occur --> - OS: PopOS - Database: PostgreSQL with PostGIS - Node.js version: v14.13.0 - Prisma version: ``` @prisma/cli : 2.13.0 @prisma/client : 2.13.0 Current platform : debian-openssl-1.1.x Query Engine : query-engine 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x) Migration Engine : migration-engine-cli 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x) Introspection Engine : introspection-core 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x) Format Binary : prisma-fmt 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x) Studio : 0.329.0 ```
1.0
unable to run migrations in an empty PostGIS db - <!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description Copied from https://github.com/prisma/prisma/discussions/4732 <!-- A clear and concise description of what the bug is. --> I'm struggling to use the migration tool after I upgraded to 2.13.0. I have a few spatial columns that I manage with scripts besides the Prisma migration tool. My pipeline used to be this: 1. apply Prisma migrations 2. run manual migrations and add / alter spatial columns if needed but with 2.13.0, I can no longer run any Prisma migration on a fresh PostGIS database because of this error: ``` Error: P3005 The database schema for `localhost:5432` is not empty. Read more about how to baseline an existing production database: https://pris.ly/d/migrate-baseline ``` Enabling PostGIS creates a `spatial_ref_sys` table, so technically the database is not empty, but it should not affect migrations. I know Prisma does not yet support the PostGIS extension, but I'm looking for a workaround. I'm also a Nexus and nexus-plugin-prisma user which doesn't support 2.12.x, so I can't downgrade reliably. ## How to reproduce Steps to reproduce the behavior: 1. Have an empty PostGIS database (I used https://hub.docker.com/r/postgis/postgis) 1. Run `prisma migrate dev --preview-feature` 1. See error ## Expected behavior Migrations run without an error. ## Environment & setup <!-- In which environment does the problem occur --> - OS: PopOS - Database: PostgreSQL with PostGIS - Node.js version: v14.13.0 - Prisma version: ``` @prisma/cli : 2.13.0 @prisma/client : 2.13.0 Current platform : debian-openssl-1.1.x Query Engine : query-engine 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x) Migration Engine : migration-engine-cli 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x) Introspection Engine : introspection-core 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x) Format Binary : prisma-fmt 833ab05d2a20e822f6736a39a27de4fc8f6b3e49 (at node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x) Studio : 0.329.0 ```
process
unable to run migrations in an empty postgis db thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description copied from i m struggling to use the migration tool after i upgraded to i have a few spatial columns that i manage with scripts besides the prisma migration tool my pipeline used to be this apply prisma migrations run manual migrations and add alter spatial columns if needed but with i can no longer run any prisma migration on a fresh postgis database because of this error error the database schema for localhost is not empty read more about how to baseline an existing production database enabling postgis creates a spatial ref sys table so technically the database is not empty but it should not affect migrations i know prisma does not yet support the postgis extension but i m looking for a workaround i m also a nexus and nexus plugin prisma user which doesn t support x so i can t downgrade reliably how to reproduce steps to reproduce the behavior have an empty postgis database i used run prisma migrate dev preview feature see error expected behavior migrations run without an error environment setup os popos database postgresql with postgis node js version prisma version prisma cli prisma client current platform debian openssl x query engine query engine at node modules prisma engines query engine debian openssl x migration engine migration engine cli at node modules prisma engines migration engine debian openssl x introspection engine introspection core at node modules prisma engines introspection engine debian openssl x format binary prisma fmt at node modules prisma engines prisma fmt debian openssl x studio
1
204,205
15,428,793,361
IssuesEvent
2021-03-06 00:51:31
comit-network/xmr-btc-swap
https://api.github.com/repos/comit-network/xmr-btc-swap
opened
Alice has problems redeeming
bug test/stagenet demo
Scenario: 1. Bob locks BTC 2. Alice locks XMR 3. Alice runs into Electrum error when redeeming 4. Bob runs into Electrum error when watching for redeem Alice's error: ``` Mar 05 12:50:46.784 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:48.689 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:50.493 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:52.302 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:54.102 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:55.907 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... ``` On Alice's (ASB) side this is a severe problem where we print to the logs about every half second spamming the logs really fast. Eventually the ASB restarts (at the moment I am not sure what causes the restart, but it could be systemd picking up the log load?). Bob's error log: ``` Monero lock tx has 10 out of 10 confirmations txid=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0 Error: Transient errors should be retried Caused by: Electrum client error ``` Lock transaction hashes where this problem was recorded (Bob): * https://blockstream.info/testnet/tx/3e305e6895211f974718d3ad83e0649d3a090a726f3773c9dc415fc9888a9e24 * https://monero-stagenet.exan.tech/search?value=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0 Investigation needed why Alice runs into this error - is it load related (i.e. we spam the public Electrum instance too much) or is it related to other problems (e.g. specific addresses, ...).
1.0
Alice has problems redeeming - Scenario: 1. Bob locks BTC 2. Alice locks XMR 3. Alice runs into Electrum error when redeeming 4. Bob runs into Electrum error when watching for redeem Alice's error: ``` Mar 05 12:50:46.784 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:48.689 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:50.493 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:52.302 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:54.102 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... Mar 05 12:50:55.907 DEBUG swap{id=3f9ed55a-99f6-4eef-959d-e32e3877efd2}: swap::bitcoin::wallet: Received protocol error "missing transaction" from Electrum, retrying... ``` On Alice's (ASB) side this is a severe problem where we print to the logs about every half second spamming the logs really fast. Eventually the ASB restarts (at the moment I am not sure what causes the restart, but it could be systemd picking up the log load?). Bob's error log: ``` Monero lock tx has 10 out of 10 confirmations txid=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0 Error: Transient errors should be retried Caused by: Electrum client error ``` Lock transaction hashes where this problem was recorded (Bob): * https://blockstream.info/testnet/tx/3e305e6895211f974718d3ad83e0649d3a090a726f3773c9dc415fc9888a9e24 * https://monero-stagenet.exan.tech/search?value=d3165221823977a2c0cadb9e5cd59a0b04c95c46ec6023dfc7a8c98f6656b0e0 Investigation needed why Alice runs into this error - is it load related (i.e. we spam the public Electrum instance too much) or is it related to other problems (e.g. specific addresses, ...).
non_process
alice has problems redeeming scenario bob locks btc alice locks xmr alice runs into electrum error when redeeming bob runs into electrum error when watching for redeem alice s error mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying mar debug swap id swap bitcoin wallet received protocol error missing transaction from electrum retrying on alice s asb side this is a severe problem where we print to the logs about every half second spamming the logs really fast eventually the asb restarts at the moment i am not sure what causes the restart but it could be systemd picking up the log load bob s error log monero lock tx has out of confirmations txid error transient errors should be retried caused by electrum client error lock transaction hashes where this problem was recorded bob investigation needed why alice runs into this error is it load related i e we spam the public electrum instance too much or is it related to other problems e g specific addresses
0
4,247
7,187,154,477
IssuesEvent
2018-02-02 03:16:18
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Monitor speed improvement
monitors-all status-inprocess type-enhancement
For non-contract accounts the only thing one needs to check to see if the block is of interest is if the account's balance changed. The only way for a non-contract account to have been involved in a transaction is if it a. sent a transaction in which case its balance changed via gas usage, b. received ether in which case its balance changed, or c. was involved in two, or four, or six, or 2*n transactions that exactly offset it in the same block in which case its balance will not have changed. This last can be found in --deep scan, but we might be able to shortcut the search if the account is not a smart contract and it's balance did not change. Question: is querying for balance worse than or faster than scanning the transactions? My guess is the former by a very large margin.
1.0
Monitor speed improvement - For non-contract accounts the only thing one needs to check to see if the block is of interest is if the account's balance changed. The only way for a non-contract account to have been involved in a transaction is if it a. sent a transaction in which case its balance changed via gas usage, b. received ether in which case its balance changed, or c. was involved in two, or four, or six, or 2*n transactions that exactly offset it in the same block in which case its balance will not have changed. This last can be found in --deep scan, but we might be able to shortcut the search if the account is not a smart contract and it's balance did not change. Question: is querying for balance worse than or faster than scanning the transactions? My guess is the former by a very large margin.
process
monitor speed improvement for non contract accounts the only thing one needs to check to see if the block is of interest is if the account s balance changed the only way for a non contract account to have been involved in a transaction is if it a sent a transaction in which case its balance changed via gas usage b received ether in which case its balance changed or c was involved in two or four or six or n transactions that exactly offset it in the same block in which case its balance will not have changed this last can be found in deep scan but we might be able to shortcut the search if the account is not a smart contract and it s balance did not change question is querying for balance worse than or faster than scanning the transactions my guess is the former by a very large margin
1
6,458
9,546,567,737
IssuesEvent
2019-05-01 20:19:41
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Department of State: Transcripts
Apply Process Approved Requirements Ready State Dept.
Who: Student Applicant What: Transcript page Why: The student is required to submit a transcript with their application. - There will be a header "Transcripts" - There will be a list of transcripts the user has stored in their USAJOBS profile - The user can select the transcript they would like to submit with their application - The "Upload transcript" link will take the user to their USAJOBS profile to upload a transcript (this will open in a new window) - When the user returns to their application there will be a "refresh transcripts" button to update their application (this was discussed in our sprint planning meeting- because the transcripts are uploaded in USAJOBS we need to pull the new transcript into the application) - When the user clicks the "view" button it will act the same as it does on USAJOBS (this was discussed during sprint planning- the view functioanlity will be dependent on the users browser) - The "Back" Button will take the user to the "Experience & References" page - The "Save and continue" button will take the user to the "Languages & Skills" page. - New Transcripts are added and saved on USAJOBS. Public Link: https://opm.invisionapp.com/share/ZEPNZR09Q54
1.0
Department of State: Transcripts - Who: Student Applicant What: Transcript page Why: The student is required to submit a transcript with their application. - There will be a header "Transcripts" - There will be a list of transcripts the user has stored in their USAJOBS profile - The user can select the transcript they would like to submit with their application - The "Upload transcript" link will take the user to their USAJOBS profile to upload a transcript (this will open in a new window) - When the user returns to their application there will be a "refresh transcripts" button to update their application (this was discussed in our sprint planning meeting- because the transcripts are uploaded in USAJOBS we need to pull the new transcript into the application) - When the user clicks the "view" button it will act the same as it does on USAJOBS (this was discussed during sprint planning- the view functioanlity will be dependent on the users browser) - The "Back" Button will take the user to the "Experience & References" page - The "Save and continue" button will take the user to the "Languages & Skills" page. - New Transcripts are added and saved on USAJOBS. Public Link: https://opm.invisionapp.com/share/ZEPNZR09Q54
process
department of state transcripts who student applicant what transcript page why the student is required to submit a transcript with their application there will be a header transcripts there will be a list of transcripts the user has stored in their usajobs profile the user can select the transcript they would like to submit with their application the upload transcript link will take the user to their usajobs profile to upload a transcript this will open in a new window when the user returns to their application there will be a refresh transcripts button to update their application this was discussed in our sprint planning meeting because the transcripts are uploaded in usajobs we need to pull the new transcript into the application when the user clicks the view button it will act the same as it does on usajobs this was discussed during sprint planning the view functioanlity will be dependent on the users browser the back button will take the user to the experience references page the save and continue button will take the user to the languages skills page new transcripts are added and saved on usajobs public link
1
39,369
9,416,542,584
IssuesEvent
2019-04-10 14:52:43
SasView/sasview
https://api.github.com/repos/SasView/sasview
opened
5.0 plugin product model error ?
defect
In my local build of 5.0, ESS_GUI, I made a plugin model for sphere times hardsphere This does not work - likely a feature not yet coded - see error message below. ALSO need to decide how we handle the beta(Q) options, structure_factor_mode, radius_effective_mode etc. 15:43:17 - INFO: 2019-04-10 15:43:17 15:43:17 - ERROR: Traceback (most recent call last): File "C:\sasview42\sasview\src\sas\sascalc\data_util\calcthread.py", line 274, in _run self.compute(*args, **kwargs) File "C:\sasview42\sasview\src\sas\qtgui\Perspectives\Fitting\ModelThread.py", line 211, in compute intermediate_results = intermediate_results() TypeError: 'list' object is not callable
1.0
5.0 plugin product model error ? - In my local build of 5.0, ESS_GUI, I made a plugin model for sphere times hardsphere This does not work - likely a feature not yet coded - see error message below. ALSO need to decide how we handle the beta(Q) options, structure_factor_mode, radius_effective_mode etc. 15:43:17 - INFO: 2019-04-10 15:43:17 15:43:17 - ERROR: Traceback (most recent call last): File "C:\sasview42\sasview\src\sas\sascalc\data_util\calcthread.py", line 274, in _run self.compute(*args, **kwargs) File "C:\sasview42\sasview\src\sas\qtgui\Perspectives\Fitting\ModelThread.py", line 211, in compute intermediate_results = intermediate_results() TypeError: 'list' object is not callable
non_process
plugin product model error in my local build of ess gui i made a plugin model for sphere times hardsphere this does not work likely a feature not yet coded see error message below also need to decide how we handle the beta q options structure factor mode radius effective mode etc info error traceback most recent call last file c sasview src sas sascalc data util calcthread py line in run self compute args kwargs file c sasview src sas qtgui perspectives fitting modelthread py line in compute intermediate results intermediate results typeerror list object is not callable
0
6,502
9,576,150,108
IssuesEvent
2019-05-07 08:25:09
DCRGraphsNet/DCROpenCaseManager
https://api.github.com/repos/DCRGraphsNet/DCROpenCaseManager
closed
As a social worker I should be able to see the status of single processes of a child. (3)
Must have Page: Child Page: Process Sprint5
What is the status actually (ask Syddjurs)? Status of each process of a child. How far in the process are we? Show Green (all ok) or Red (deadline broken)
1.0
As a social worker I should be able to see the status of single processes of a child. (3) - What is the status actually (ask Syddjurs)? Status of each process of a child. How far in the process are we? Show Green (all ok) or Red (deadline broken)
process
as a social worker i should be able to see the status of single processes of a child what is the status actually ask syddjurs status of each process of a child how far in the process are we show green all ok or red deadline broken
1
93,355
26,933,048,691
IssuesEvent
2023-02-07 18:15:36
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Add support for VS2022 build chain
module: build module: windows triaged
**Requirements** - PyTorch can be compiled with VS2022 build chain. **Acceptance Criteria** - CD pipelines (`ciflow-binaries`) are producing binaries using VS 2022. cc @malfet @seemethere @peterjc123 @mszhanyi @skyline75489 @nbcsm
1.0
Add support for VS2022 build chain - **Requirements** - PyTorch can be compiled with VS2022 build chain. **Acceptance Criteria** - CD pipelines (`ciflow-binaries`) are producing binaries using VS 2022. cc @malfet @seemethere @peterjc123 @mszhanyi @skyline75489 @nbcsm
non_process
add support for build chain requirements pytorch can be compiled with build chain acceptance criteria cd pipelines ciflow binaries are producing binaries using vs cc malfet seemethere mszhanyi nbcsm
0
14,143
17,034,746,346
IssuesEvent
2021-07-05 04:51:34
theislab/scanpy
https://api.github.com/repos/theislab/scanpy
opened
Better handling of upstream releases
Development Process 🚀 Enhancement ✨
It would be nice if we had a better way of handling upstream releases. E.g. when pandas makes a release it would be good that we had tested against their release candidates, or if we had a good process for dealing with bugs if they do occur. One think we could do, is defensively pin dependencies to below their current release series. I don't like doing this since I think it's pretty restrictive when most of the time we don't have issues. Maybe we could do this for breaking releases, but that wouldn't have prevented issues like #1917. It would be nice to automate the process of testing against upstream release candidates. Basically, when something comes out, we build against it so we can report issues early and don't have to deal with it in live releases. I'm not sure how to do this with `pip search` not working anymore.
1.0
Better handling of upstream releases - It would be nice if we had a better way of handling upstream releases. E.g. when pandas makes a release it would be good that we had tested against their release candidates, or if we had a good process for dealing with bugs if they do occur. One think we could do, is defensively pin dependencies to below their current release series. I don't like doing this since I think it's pretty restrictive when most of the time we don't have issues. Maybe we could do this for breaking releases, but that wouldn't have prevented issues like #1917. It would be nice to automate the process of testing against upstream release candidates. Basically, when something comes out, we build against it so we can report issues early and don't have to deal with it in live releases. I'm not sure how to do this with `pip search` not working anymore.
process
better handling of upstream releases it would be nice if we had a better way of handling upstream releases e g when pandas makes a release it would be good that we had tested against their release candidates or if we had a good process for dealing with bugs if they do occur one think we could do is defensively pin dependencies to below their current release series i don t like doing this since i think it s pretty restrictive when most of the time we don t have issues maybe we could do this for breaking releases but that wouldn t have prevented issues like it would be nice to automate the process of testing against upstream release candidates basically when something comes out we build against it so we can report issues early and don t have to deal with it in live releases i m not sure how to do this with pip search not working anymore
1
1,709
3,897,596,180
IssuesEvent
2016-04-16 14:41:22
justarrived/just_match_api
https://api.github.com/repos/justarrived/just_match_api
closed
Frilans Finans invoice implementation
3rd-party-service in progress must have
Currently the external API is not live yet and too unstable to mock.
1.0
Frilans Finans invoice implementation - Currently the external API is not live yet and too unstable to mock.
non_process
frilans finans invoice implementation currently the external api is not live yet and too unstable to mock
0
412,726
27,869,434,476
IssuesEvent
2023-03-21 12:33:51
s1dlx/sd-webui-bayesian-merger
https://api.github.com/repos/s1dlx/sd-webui-bayesian-merger
closed
Need a brief description for --scorer_model_dir
documentation
First off Awesome work! Its lacking a quick sentence about how it deals with the scorer models. Not sure if it tries all models or can even load more than one. I copied all the models from the https://github.com/Xerxemi/sdweb-auto-MBW plugin. Are there others that work with it is that even a good idea? Im assuming this option is not needed if I just drop models into that folder?? Thanks for all your hard work!
1.0
Need a brief description for --scorer_model_dir - First off Awesome work! Its lacking a quick sentence about how it deals with the scorer models. Not sure if it tries all models or can even load more than one. I copied all the models from the https://github.com/Xerxemi/sdweb-auto-MBW plugin. Are there others that work with it is that even a good idea? Im assuming this option is not needed if I just drop models into that folder?? Thanks for all your hard work!
non_process
need a brief description for scorer model dir first off awesome work its lacking a quick sentence about how it deals with the scorer models not sure if it tries all models or can even load more than one i copied all the models from the plugin are there others that work with it is that even a good idea im assuming this option is not needed if i just drop models into that folder thanks for all your hard work
0