Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
21,367
| 29,194,080,693
|
IssuesEvent
|
2023-05-20 00:31:50
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] QA Analyst na Coodesh
|
SALVADOR TESTE PJ JAVASCRIPT RUBY REQUISITOS SELENIUM CUCUMBER REMOTO CYPRESS PROCESSOS GITHUB UMA QUALIDADE VENDAS QA TESTES AUTOMATIZADOS METODOLOGIAS ÁGEIS RSPEC AUTOMAÇÃO DE TESTES Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>BaladAPP </strong>está em busca de <strong><ins>QA Analyst</ins></strong> para compor seu time!<br><br>A BaladAPP é uma das maiores ticketeiras do Brasil, atuando de norte a sul, atendendo mais de 200 mil clientes mensalmente. <br>Buscamos sempre trazer soluções para produtores de eventos auxiliando desde o início das vendas online até o dia do evento. <br><br><strong>Responsabilidades:</strong></p>
<ul>
<li>Estamos procurando por um profissional de qualidade de software para se juntar ao nosso time e estruturar nosso setor de qualidade. Como membro da equipe, você será responsável por testar nossos sistemas, fazer a gestão das nossas versões e manter nosso changelog. Além disso, você terá a oportunidade de evoluir e avaliar nossos testes automatizados, a fim de tornar nossos processos de teste mais eficientes.</li>
</ul>
<p><br><br></p>
## BALADAPP:
<p>A BaladAPP é uma empresa especializada em vendas de ingressos para eventos nacionais e internacionais, focados em shows e espetáculos, funcionando como marketplace, trabalhamos com eventos de pequeno à grande porte.</p>
</p>
## Habilidades:
- RSpec
- Cucumber
- Cypress
- Selenium
- Automação de Testes
## Local:
100% Remoto
## Requisitos:
- Experiência anterior como Analista de Testes/QA;
- Experiência com testes automatizados com uma destas ferramentas Selenium, Cypress, Cucumber (Preferencialmente com RSpec);
- Conhecimento em Ruby;
- Conhecimentos em processos e metodologias ágeis;
- Habilidades de resolução de problemas e pensamento crítico.
## Diferenciais:
- Conhecimento de linguagens de programação Javascript.
## Benefícios:
- Horários flexíveis;
- 30 dias de férias remuneradas no PJ;
- Plano de Saúde;
- Plano Odontológico.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [QA Analyst na BALADAPP](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Testes/Q.A
|
1.0
|
[Remoto] QA Analyst na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>BaladAPP </strong>está em busca de <strong><ins>QA Analyst</ins></strong> para compor seu time!<br><br>A BaladAPP é uma das maiores ticketeiras do Brasil, atuando de norte a sul, atendendo mais de 200 mil clientes mensalmente. <br>Buscamos sempre trazer soluções para produtores de eventos auxiliando desde o início das vendas online até o dia do evento. <br><br><strong>Responsabilidades:</strong></p>
<ul>
<li>Estamos procurando por um profissional de qualidade de software para se juntar ao nosso time e estruturar nosso setor de qualidade. Como membro da equipe, você será responsável por testar nossos sistemas, fazer a gestão das nossas versões e manter nosso changelog. Além disso, você terá a oportunidade de evoluir e avaliar nossos testes automatizados, a fim de tornar nossos processos de teste mais eficientes.</li>
</ul>
<p><br><br></p>
## BALADAPP:
<p>A BaladAPP é uma empresa especializada em vendas de ingressos para eventos nacionais e internacionais, focados em shows e espetáculos, funcionando como marketplace, trabalhamos com eventos de pequeno à grande porte.</p>
</p>
## Habilidades:
- RSpec
- Cucumber
- Cypress
- Selenium
- Automação de Testes
## Local:
100% Remoto
## Requisitos:
- Experiência anterior como Analista de Testes/QA;
- Experiência com testes automatizados com uma destas ferramentas Selenium, Cypress, Cucumber (Preferencialmente com RSpec);
- Conhecimento em Ruby;
- Conhecimentos em processos e metodologias ágeis;
- Habilidades de resolução de problemas e pensamento crítico.
## Diferenciais:
- Conhecimento de linguagens de programação Javascript.
## Benefícios:
- Horários flexíveis;
- 30 dias de férias remuneradas no PJ;
- Plano de Saúde;
- Plano Odontológico.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [QA Analyst na BALADAPP](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Testes/Q.A
|
process
|
qa analyst na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a baladapp está em busca de qa analyst para compor seu time a baladapp é uma das maiores ticketeiras do brasil atuando de norte a sul atendendo mais de mil clientes mensalmente buscamos sempre trazer soluções para produtores de eventos auxiliando desde o início das vendas online até o dia do evento responsabilidades estamos procurando por um profissional de qualidade de software para se juntar ao nosso time e estruturar nosso setor de qualidade como membro da equipe você será responsável por testar nossos sistemas fazer a gestão das nossas versões e manter nosso changelog além disso você terá a oportunidade de evoluir e avaliar nossos testes automatizados a fim de tornar nossos processos de teste mais eficientes baladapp a baladapp é uma empresa especializada em vendas de ingressos para eventos nacionais e internacionais focados em shows e espetáculos funcionando como marketplace trabalhamos com eventos de pequeno à grande porte habilidades rspec cucumber cypress selenium automação de testes local remoto requisitos experiência anterior como analista de testes qa experiência com testes automatizados com uma destas ferramentas selenium cypress cucumber preferencialmente com rspec conhecimento em ruby conhecimentos em processos e metodologias ágeis habilidades de resolução de problemas e pensamento crítico diferenciais conhecimento de linguagens de programação javascript benefícios horários flexíveis dias de férias remuneradas no pj plano de saúde plano odontológico como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria testes q a
| 1
|
7,259
| 10,420,471,777
|
IssuesEvent
|
2019-09-16 00:42:00
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
In place processing multipart to singleparts does not handle unique constraints
|
Bug Processing
|
Provider errors are generated due to failure of UNIQUE constraint when you try to save after running Multipart to singleparts in a geopackage layer.
**How to Reproduce**
1. Download and open a copy of ‘Multi_to_Singleparts_FID_Bug.qgz’ (zip attached below)
2. Open the Attribute Table for the ‘Multi_to_Singleparts_FID_Bug’ layer.
3. Select the multipart feature with FID value of 4.
4. Ctrl + K
5. “ef multipart to singleparts”
6. Observe there are now 5 features selected in the attribute table with FID value of 4 (This is a bug).
7. Undo, and the edit is successfully reversed.
8. Repeat steps 3 to 6.
9. Click ‘Save Layer Edits’
10. Observe there are 6 Commit errors, including:
Could not commit changes to layer Multi_to_Singleparts_FID_Bug
Errors: SUCCESS: 1 feature(s) deleted.
ERROR: 5 feature(s) not added.
Provider errors:
OGR error creating feature -8: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
OGR error creating feature -9: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
OGR error creating feature -10: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
OGR error creating feature -11: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
11. On the Map Canvas, marquee select the feature(s) at the south-west corner.
12. In the attribute table, observe that the title bar reports 2 features selected, but there is only one feature selected in the table (bug in attribute table failing to update).
13. Close and re-open the attribute table, and observe there are now 2 features selected out of 6 features with FID value of 4.
14. Either ‘Undo’ or ‘Toggle Edit & Discard Changes’.
15. Observe that 5 of the 6 FID value-4 features have been deleted, leaving only one remaining.
Workaround:
16. Repeat steps 1 to 6
17. In the attribute table, with the relevant features selected, change the FID value to null.
18. Save Layer Edits.
19. Observe that unique values have now been autogenerated.
**QGIS and OS versions**
QGIS 3.4.11 (LTR), 3.8.2, and 3.9 (nightly build)
[Multi_to_Singleparts_FID_Bug.zip](https://github.com/qgis/QGIS/files/3590496/Multi_to_Singleparts_FID_Bug.zip)
|
1.0
|
In place processing multipart to singleparts does not handle unique constraints - Provider errors are generated due to failure of UNIQUE constraint when you try to save after running Multipart to singleparts in a geopackage layer.
**How to Reproduce**
1. Download and open a copy of ‘Multi_to_Singleparts_FID_Bug.qgz’ (zip attached below)
2. Open the Attribute Table for the ‘Multi_to_Singleparts_FID_Bug’ layer.
3. Select the multipart feature with FID value of 4.
4. Ctrl + K
5. “ef multipart to singleparts”
6. Observe there are now 5 features selected in the attribute table with FID value of 4 (This is a bug).
7. Undo, and the edit is successfully reversed.
8. Repeat steps 3 to 6.
9. Click ‘Save Layer Edits’
10. Observe there are 6 Commit errors, including:
Could not commit changes to layer Multi_to_Singleparts_FID_Bug
Errors: SUCCESS: 1 feature(s) deleted.
ERROR: 5 feature(s) not added.
Provider errors:
OGR error creating feature -8: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
OGR error creating feature -9: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
OGR error creating feature -10: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
OGR error creating feature -11: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid
11. On the Map Canvas, marquee select the feature(s) at the south-west corner.
12. In the attribute table, observe that the title bar reports 2 features selected, but there is only one feature selected in the table (bug in attribute table failing to update).
13. Close and re-open the attribute table, and observe there are now 2 features selected out of 6 features with FID value of 4.
14. Either ‘Undo’ or ‘Toggle Edit & Discard Changes’.
15. Observe that 5 of the 6 FID value-4 features have been deleted, leaving only one remaining.
Workaround:
16. Repeat steps 1 to 6
17. In the attribute table, with the relevant features selected, change the FID value to null.
18. Save Layer Edits.
19. Observe that unique values have now been autogenerated.
**QGIS and OS versions**
QGIS 3.4.11 (LTR), 3.8.2, and 3.9 (nightly build)
[Multi_to_Singleparts_FID_Bug.zip](https://github.com/qgis/QGIS/files/3590496/Multi_to_Singleparts_FID_Bug.zip)
|
process
|
in place processing multipart to singleparts does not handle unique constraints provider errors are generated due to failure of unique constraint when you try to save after running multipart to singleparts in a geopackage layer how to reproduce download and open a copy of ‘multi to singleparts fid bug qgz’ zip attached below open the attribute table for the ‘multi to singleparts fid bug’ layer select the multipart feature with fid value of ctrl k “ef multipart to singleparts” observe there are now features selected in the attribute table with fid value of this is a bug undo and the edit is successfully reversed repeat steps to click ‘save layer edits’ observe there are commit errors including could not commit changes to layer multi to singleparts fid bug errors success feature s deleted error feature s not added provider errors ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid on the map canvas marquee select the feature s at the south west corner in the attribute table observe that the title bar reports features selected but there is only one feature selected in the table bug in attribute table failing to update close and re open the attribute table and observe there are now features selected out of features with fid value of either ‘undo’ or ‘toggle edit discard changes’ observe that of the fid value features have been deleted leaving only one remaining workaround repeat steps to in the attribute table with the relevant features selected change the fid value to null save layer edits observe that unique values have now been autogenerated qgis and os versions qgis ltr and nightly build
| 1
|
71,721
| 3,367,617,921
|
IssuesEvent
|
2015-11-22 10:19:05
|
music-encoding/music-encoding
|
https://api.github.com/repos/music-encoding/music-encoding
|
closed
|
The element head is not allowed in projectDesc
|
Priority: Medium
|
_From [siggelun...@gmail.com](https://code.google.com/u/110461478002540803867/) on December 19, 2013 03:06:13_
What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? What version of the product are you using? On what operating system? Please provide any additional information below. In my view \<projectDesc> and elements with similar function should have a more fully fledge support for prose.
_Original issue: http://code.google.com/p/music-encoding/issues/detail?id=187_
|
1.0
|
The element head is not allowed in projectDesc - _From [siggelun...@gmail.com](https://code.google.com/u/110461478002540803867/) on December 19, 2013 03:06:13_
What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? What version of the product are you using? On what operating system? Please provide any additional information below. In my view \<projectDesc> and elements with similar function should have a more fully fledge support for prose.
_Original issue: http://code.google.com/p/music-encoding/issues/detail?id=187_
|
non_process
|
the element head is not allowed in projectdesc from on december what steps will reproduce the problem what is the expected output what do you see instead what version of the product are you using on what operating system please provide any additional information below in my view and elements with similar function should have a more fully fledge support for prose original issue
| 0
|
17,141
| 22,686,040,208
|
IssuesEvent
|
2022-07-04 14:11:13
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Enhance randomized process tests with start process instance anywhere ability
|
team/process-automation
|
To comprehensively test the start process instance anywhere feature, we'll need to enhance the randomized process tests with the ability to start the process instance anywhere.
Blocked by #9390 #9391
## Out of scope
- starting the process instance inside a multi-instance
- starting the process instance inside a call activity
|
1.0
|
Enhance randomized process tests with start process instance anywhere ability - To comprehensively test the start process instance anywhere feature, we'll need to enhance the randomized process tests with the ability to start the process instance anywhere.
Blocked by #9390 #9391
## Out of scope
- starting the process instance inside a multi-instance
- starting the process instance inside a call activity
|
process
|
enhance randomized process tests with start process instance anywhere ability to comprehensively test the start process instance anywhere feature we ll need to enhance the randomized process tests with the ability to start the process instance anywhere blocked by out of scope starting the process instance inside a multi instance starting the process instance inside a call activity
| 1
|
7,406
| 10,525,900,731
|
IssuesEvent
|
2019-09-30 15:56:15
|
Python-Markdown/markdown
|
https://api.github.com/repos/Python-Markdown/markdown
|
reopened
|
Python Version Support Timeline
|
process
|
Just noting this so I can find it later. Python defines the status of its versions [here][1].
The following end-of life cycles are currently scheduled:
* <del>Python 3.4 2019-03-16</del>
* Python 2.7 2020-01-01
* Python 3.5 2020-09-13
* Python 3.6 2021-12-23
* Python 3.7 2023-06-27
* Python 3.8 2024-10
Also of interest is https://python3statement.org/ where various projects have signed on to a pledge to discontinue support for Python 2,7 on or before 2020. I'm indifferent to joining the pledge, but the arguments given for dropping support could be useful when we take that move.
I intend to keep this issue open indefinitely to track future updates. As a general rule, I expect we will drop support for a Python version with our next minor release after Python has dropped support. That said, if/when Travis drops support, our tests stop passing, so we have occasionally dropped support quickly in the past. Of course, as per our Contributing Guidelines, such a change should not happen in a bugfix release.
[1]: https://devguide.python.org/#status-of-python-branches
|
1.0
|
Python Version Support Timeline - Just noting this so I can find it later. Python defines the status of its versions [here][1].
The following end-of life cycles are currently scheduled:
* <del>Python 3.4 2019-03-16</del>
* Python 2.7 2020-01-01
* Python 3.5 2020-09-13
* Python 3.6 2021-12-23
* Python 3.7 2023-06-27
* Python 3.8 2024-10
Also of interest is https://python3statement.org/ where various projects have signed on to a pledge to discontinue support for Python 2,7 on or before 2020. I'm indifferent to joining the pledge, but the arguments given for dropping support could be useful when we take that move.
I intend to keep this issue open indefinitely to track future updates. As a general rule, I expect we will drop support for a Python version with our next minor release after Python has dropped support. That said, if/when Travis drops support, our tests stop passing, so we have occasionally dropped support quickly in the past. Of course, as per our Contributing Guidelines, such a change should not happen in a bugfix release.
[1]: https://devguide.python.org/#status-of-python-branches
|
process
|
python version support timeline just noting this so i can find it later python defines the status of its versions the following end of life cycles are currently scheduled python python python python python python also of interest is where various projects have signed on to a pledge to discontinue support for python on or before i m indifferent to joining the pledge but the arguments given for dropping support could be useful when we take that move i intend to keep this issue open indefinitely to track future updates as a general rule i expect we will drop support for a python version with our next minor release after python has dropped support that said if when travis drops support our tests stop passing so we have occasionally dropped support quickly in the past of course as per our contributing guidelines such a change should not happen in a bugfix release
| 1
|
18,085
| 24,107,887,533
|
IssuesEvent
|
2022-09-20 08:58:04
|
googleapis/dotnet-spanner-nhibernate
|
https://api.github.com/repos/googleapis/dotnet-spanner-nhibernate
|
closed
|
Dependency Dashboard
|
type: process api: spanner
|
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/benchmarkdotnet-0.x -->[chore(deps): update dependency benchmarkdotnet to v0.13.2](../pull/98)
- [ ] <!-- rebase-branch=renovate/docker.dotnet-3.x -->[chore(deps): update dependency docker.dotnet to v3.125.12](../pull/102)
- [ ] <!-- rebase-branch=renovate/gcp-releasetool-1.x -->[chore(deps): update dependency gcp-releasetool to v1.8.8](../pull/109)
- [ ] <!-- rebase-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2.11.1](../pull/112)
- [ ] <!-- rebase-branch=renovate/keyring-23.x -->[chore(deps): update dependency keyring to v23.9.3](../pull/111)
- [ ] <!-- rebase-branch=renovate/vstest-monorepo -->[chore(deps): update dependency microsoft.net.test.sdk to v17.3.1](../pull/99)
- [ ] <!-- rebase-branch=renovate/nhibernate-5.x -->[chore(deps): update dependency nhibernate to v5.3.13](../pull/101)
- [ ] <!-- rebase-branch=renovate/pyjwt-2.x -->[chore(deps): update dependency pyjwt to v2.5.0](../pull/110)
- [ ] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once**
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/certifi-2022.x -->[chore(deps): update dependency certifi to v2022.9.14](../pull/106)
- [ ] <!-- recreate-branch=renovate/click-8.x -->[chore(deps): update dependency click to v8.1.3](../pull/104)
- [ ] <!-- recreate-branch=renovate/google.cloud.spanner.data-4.x -->[chore(deps): update dependency google.cloud.spanner.data to v4](../pull/94)
- [ ] <!-- recreate-branch=renovate/protobuf-4.x -->[chore(deps): update dependency protobuf to v4](../pull/105)
## Detected dependencies
<details><summary>github-actions</summary>
<blockquote>
<details><summary>.github/workflows/build-and-test-on-windows.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
<details><summary>.github/workflows/build-and-test.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
<details><summary>.github/workflows/integration-tests-on-emulator.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
<details><summary>.github/workflows/integration-tests-on-production.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
</blockquote>
</details>
<details><summary>nuget</summary>
<blockquote>
<details><summary>Google.Cloud.Spanner.Connection.Tests/Google.Cloud.Spanner.Connection.Tests.csproj</summary>
- `coverlet.collector 3.1.2`
- `xunit.runner.visualstudio 2.4.5`
- `Xunit.Combinatorial 1.4.1`
- `xunit 2.4.2`
- `Microsoft.NET.Test.Sdk 17.3.0`
</details>
<details><summary>Google.Cloud.Spanner.Connection/Google.Cloud.Spanner.Connection.csproj</summary>
- `Google.Cloud.Spanner.Data 3.15.1`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.Benchmarks/Google.Cloud.Spanner.NHibernate.Benchmarks.csproj</summary>
- `BenchmarkDotNet 0.13.1`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.IntegrationTests/Google.Cloud.Spanner.NHibernate.IntegrationTests.csproj</summary>
- `Xunit.SkippableFact 1.4.13`
- `coverlet.collector 3.1.2`
- `xunit.runner.visualstudio 2.4.5`
- `Xunit.Combinatorial 1.4.1`
- `xunit 2.4.2`
- `Microsoft.NET.Test.Sdk 17.3.0`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.Samples/Google.Cloud.Spanner.NHibernate.Samples.csproj</summary>
- `Docker.DotNet 3.125.10`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.Tests/Google.Cloud.Spanner.NHibernate.Tests.csproj</summary>
- `coverlet.collector 3.1.2`
- `xunit.runner.visualstudio 2.4.5`
- `Xunit.Combinatorial 1.4.1`
- `xunit 2.4.2`
- `Microsoft.NET.Test.Sdk 17.3.0`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate/Google.Cloud.Spanner.NHibernate.csproj</summary>
- `NHibernate 5.3.12`
</details>
</blockquote>
</details>
<details><summary>pip_requirements</summary>
<blockquote>
<details><summary>.kokoro/requirements.txt</summary>
- `attrs ==22.1.0`
- `cachetools ==5.2.0`
- `certifi ==2022.6.15.2`
- `cffi ==1.15.1`
- `charset-normalizer ==2.1.1`
- `click ==8.0.4`
- `colorama ==0.4.5`
- `cryptography ==38.0.1`
- `gcp-releasetool ==1.8.7`
- `google-auth ==2.11.0`
- `idna ==3.4`
- `importlib-metadata ==4.12.0`
- `jaraco-classes ==3.2.2`
- `jinja2 ==3.1.2`
- `keyring ==23.9.1`
- `markupsafe ==2.1.1`
- `more-itertools ==8.14.0`
- `packaging ==21.3`
- `protobuf ==3.20.2`
- `pyasn1 ==0.4.8`
- `pyasn1-modules ==0.2.8`
- `pycparser ==2.21`
- `pyjwt ==2.4.0`
- `pyparsing ==3.0.9`
- `pyperclip ==1.8.2`
- `python-dateutil ==2.8.2`
- `pywin32-ctypes ==0.2.0`
- `requests ==2.28.1`
- `rsa ==4.9`
- `six ==1.16.0`
- `typing-extensions ==4.3.0`
- `urllib3 ==1.26.12`
- `zipp ==3.8.1`
</details>
</blockquote>
</details>
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/benchmarkdotnet-0.x -->[chore(deps): update dependency benchmarkdotnet to v0.13.2](../pull/98)
- [ ] <!-- rebase-branch=renovate/docker.dotnet-3.x -->[chore(deps): update dependency docker.dotnet to v3.125.12](../pull/102)
- [ ] <!-- rebase-branch=renovate/gcp-releasetool-1.x -->[chore(deps): update dependency gcp-releasetool to v1.8.8](../pull/109)
- [ ] <!-- rebase-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2.11.1](../pull/112)
- [ ] <!-- rebase-branch=renovate/keyring-23.x -->[chore(deps): update dependency keyring to v23.9.3](../pull/111)
- [ ] <!-- rebase-branch=renovate/vstest-monorepo -->[chore(deps): update dependency microsoft.net.test.sdk to v17.3.1](../pull/99)
- [ ] <!-- rebase-branch=renovate/nhibernate-5.x -->[chore(deps): update dependency nhibernate to v5.3.13](../pull/101)
- [ ] <!-- rebase-branch=renovate/pyjwt-2.x -->[chore(deps): update dependency pyjwt to v2.5.0](../pull/110)
- [ ] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once**
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/certifi-2022.x -->[chore(deps): update dependency certifi to v2022.9.14](../pull/106)
- [ ] <!-- recreate-branch=renovate/click-8.x -->[chore(deps): update dependency click to v8.1.3](../pull/104)
- [ ] <!-- recreate-branch=renovate/google.cloud.spanner.data-4.x -->[chore(deps): update dependency google.cloud.spanner.data to v4](../pull/94)
- [ ] <!-- recreate-branch=renovate/protobuf-4.x -->[chore(deps): update dependency protobuf to v4](../pull/105)
## Detected dependencies
<details><summary>github-actions</summary>
<blockquote>
<details><summary>.github/workflows/build-and-test-on-windows.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
<details><summary>.github/workflows/build-and-test.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
<details><summary>.github/workflows/integration-tests-on-emulator.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
<details><summary>.github/workflows/integration-tests-on-production.yml</summary>
- `actions/checkout v3`
- `actions/setup-dotnet v2`
</details>
</blockquote>
</details>
<details><summary>nuget</summary>
<blockquote>
<details><summary>Google.Cloud.Spanner.Connection.Tests/Google.Cloud.Spanner.Connection.Tests.csproj</summary>
- `coverlet.collector 3.1.2`
- `xunit.runner.visualstudio 2.4.5`
- `Xunit.Combinatorial 1.4.1`
- `xunit 2.4.2`
- `Microsoft.NET.Test.Sdk 17.3.0`
</details>
<details><summary>Google.Cloud.Spanner.Connection/Google.Cloud.Spanner.Connection.csproj</summary>
- `Google.Cloud.Spanner.Data 3.15.1`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.Benchmarks/Google.Cloud.Spanner.NHibernate.Benchmarks.csproj</summary>
- `BenchmarkDotNet 0.13.1`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.IntegrationTests/Google.Cloud.Spanner.NHibernate.IntegrationTests.csproj</summary>
- `Xunit.SkippableFact 1.4.13`
- `coverlet.collector 3.1.2`
- `xunit.runner.visualstudio 2.4.5`
- `Xunit.Combinatorial 1.4.1`
- `xunit 2.4.2`
- `Microsoft.NET.Test.Sdk 17.3.0`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.Samples/Google.Cloud.Spanner.NHibernate.Samples.csproj</summary>
- `Docker.DotNet 3.125.10`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate.Tests/Google.Cloud.Spanner.NHibernate.Tests.csproj</summary>
- `coverlet.collector 3.1.2`
- `xunit.runner.visualstudio 2.4.5`
- `Xunit.Combinatorial 1.4.1`
- `xunit 2.4.2`
- `Microsoft.NET.Test.Sdk 17.3.0`
</details>
<details><summary>Google.Cloud.Spanner.NHibernate/Google.Cloud.Spanner.NHibernate.csproj</summary>
- `NHibernate 5.3.12`
</details>
</blockquote>
</details>
<details><summary>pip_requirements</summary>
<blockquote>
<details><summary>.kokoro/requirements.txt</summary>
- `attrs ==22.1.0`
- `cachetools ==5.2.0`
- `certifi ==2022.6.15.2`
- `cffi ==1.15.1`
- `charset-normalizer ==2.1.1`
- `click ==8.0.4`
- `colorama ==0.4.5`
- `cryptography ==38.0.1`
- `gcp-releasetool ==1.8.7`
- `google-auth ==2.11.0`
- `idna ==3.4`
- `importlib-metadata ==4.12.0`
- `jaraco-classes ==3.2.2`
- `jinja2 ==3.1.2`
- `keyring ==23.9.1`
- `markupsafe ==2.1.1`
- `more-itertools ==8.14.0`
- `packaging ==21.3`
- `protobuf ==3.20.2`
- `pyasn1 ==0.4.8`
- `pyasn1-modules ==0.2.8`
- `pycparser ==2.21`
- `pyjwt ==2.4.0`
- `pyparsing ==3.0.9`
- `pyperclip ==1.8.2`
- `python-dateutil ==2.8.2`
- `pywin32-ctypes ==0.2.0`
- `requests ==2.28.1`
- `rsa ==4.9`
- `six ==1.16.0`
- `typing-extensions ==4.3.0`
- `urllib3 ==1.26.12`
- `zipp ==3.8.1`
</details>
</blockquote>
</details>
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull pull pull pull pull pull pull click on this checkbox to rebase all open prs at once ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull pull pull detected dependencies github actions github workflows build and test on windows yml actions checkout actions setup dotnet github workflows build and test yml actions checkout actions setup dotnet github workflows integration tests on emulator yml actions checkout actions setup dotnet github workflows integration tests on production yml actions checkout actions setup dotnet nuget google cloud spanner connection tests google cloud spanner connection tests csproj coverlet collector xunit runner visualstudio xunit combinatorial xunit microsoft net test sdk google cloud spanner connection google cloud spanner connection csproj google cloud spanner data google cloud spanner nhibernate benchmarks google cloud spanner nhibernate benchmarks csproj benchmarkdotnet google cloud spanner nhibernate integrationtests google cloud spanner nhibernate integrationtests csproj xunit skippablefact coverlet collector xunit runner visualstudio xunit combinatorial xunit microsoft net test sdk google cloud spanner nhibernate samples google cloud spanner nhibernate samples csproj docker dotnet google cloud spanner nhibernate tests google cloud spanner nhibernate tests csproj coverlet collector xunit runner visualstudio xunit combinatorial xunit microsoft net test sdk google cloud spanner nhibernate google cloud spanner nhibernate csproj nhibernate pip requirements kokoro requirements txt attrs cachetools certifi cffi charset normalizer click colorama cryptography gcp releasetool google auth idna importlib metadata jaraco classes keyring markupsafe more itertools packaging protobuf modules pycparser pyjwt pyparsing pyperclip python dateutil ctypes requests rsa six typing extensions zipp check this box to trigger a request for renovate to run again on this repository
| 1
|
691,291
| 23,691,105,218
|
IssuesEvent
|
2022-08-29 10:52:49
|
googleapis/python-pubsub
|
https://api.github.com/repos/googleapis/python-pubsub
|
closed
|
tests.system.TestStreamingPull: test_streaming_pull_ack_deadline failed
|
api: pubsub type: bug priority: p2 flakybot: issue flakybot: flaky
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: b48a5a5dc43c95ce8c466686e568e72c583de2f4
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cc1da8e3-578e-4bf5-845a-55cc5db2e606), [Sponge](http://sponge2/cc1da8e3-578e-4bf5-845a-55cc5db2e606)
status: failed
<details><summary>Test output</summary><br><pre>self = <tests.system.TestStreamingPull object at 0x7f5a18fe8eb0>
publisher = <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010>
subscriber = <google.cloud.pubsub_v1.SubscriberClient object at 0x7f5a17738220>
project = 'precise-truck-742'
topic_path = 'projects/precise-truck-742/topics/t-1658003098583'
subscription_path = 'projects/precise-truck-742/subscriptions/s-1658003098583'
cleanup = [(<bound method PublisherClient.delete_topic of <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010>>, ()...erClient object at 0x7f5a17738220>>, (), {'subscription': 'projects/precise-truck-742/subscriptions/s-1658003098583'})]
def test_streaming_pull_ack_deadline(
self, publisher, subscriber, project, topic_path, subscription_path, cleanup
):
# Make sure the topic and subscription get deleted.
cleanup.append((publisher.delete_topic, (), {"topic": topic_path}))
cleanup.append(
(subscriber.delete_subscription, (), {"subscription": subscription_path})
)
# Create a topic and a subscription, then subscribe to the topic. This
# must happen before the messages are published.
publisher.create_topic(name=topic_path)
# Subscribe to the topic. This must happen before the messages
# are published.
subscriber.create_subscription(
name=subscription_path, topic=topic_path, ack_deadline_seconds=45
)
# publish some messages and wait for completion
_publish_messages(publisher, topic_path, batch_sizes=[2])
# subscribe to the topic
callback = StreamingPullCallback(
processing_time=13, # more than the default stream ACK deadline (10s)
resolve_at_msg_count=3, # one more than the published messages count
)
flow_control = types.FlowControl(max_messages=1)
subscription_future = subscriber.subscribe(
subscription_path, callback, flow_control=flow_control
)
# We expect to process the first two messages in 2 * 13 seconds, and
# any duplicate message that is re-sent by the backend in additional
# 13 seconds, totalling 39 seconds (+ overhead) --> if there have been
# no duplicates in 60 seconds, we can reasonably assume that there
# won't be any.
try:
callback.done_future.result(timeout=60)
except exceptions.TimeoutError:
# future timed out, because we received no excessive messages
assert sorted(callback.seen_message_ids) == [1, 2]
else:
> pytest.fail(
"Expected to receive 2 messages, but got at least {}.".format(
len(callback.seen_message_ids)
)
)
E Failed: Expected to receive 2 messages, but got at least 3.
tests/system.py:552: Failed</pre></details>
|
1.0
|
tests.system.TestStreamingPull: test_streaming_pull_ack_deadline failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: b48a5a5dc43c95ce8c466686e568e72c583de2f4
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cc1da8e3-578e-4bf5-845a-55cc5db2e606), [Sponge](http://sponge2/cc1da8e3-578e-4bf5-845a-55cc5db2e606)
status: failed
<details><summary>Test output</summary><br><pre>self = <tests.system.TestStreamingPull object at 0x7f5a18fe8eb0>
publisher = <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010>
subscriber = <google.cloud.pubsub_v1.SubscriberClient object at 0x7f5a17738220>
project = 'precise-truck-742'
topic_path = 'projects/precise-truck-742/topics/t-1658003098583'
subscription_path = 'projects/precise-truck-742/subscriptions/s-1658003098583'
cleanup = [(<bound method PublisherClient.delete_topic of <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010>>, ()...erClient object at 0x7f5a17738220>>, (), {'subscription': 'projects/precise-truck-742/subscriptions/s-1658003098583'})]
def test_streaming_pull_ack_deadline(
self, publisher, subscriber, project, topic_path, subscription_path, cleanup
):
# Make sure the topic and subscription get deleted.
cleanup.append((publisher.delete_topic, (), {"topic": topic_path}))
cleanup.append(
(subscriber.delete_subscription, (), {"subscription": subscription_path})
)
# Create a topic and a subscription, then subscribe to the topic. This
# must happen before the messages are published.
publisher.create_topic(name=topic_path)
# Subscribe to the topic. This must happen before the messages
# are published.
subscriber.create_subscription(
name=subscription_path, topic=topic_path, ack_deadline_seconds=45
)
# publish some messages and wait for completion
_publish_messages(publisher, topic_path, batch_sizes=[2])
# subscribe to the topic
callback = StreamingPullCallback(
processing_time=13, # more than the default stream ACK deadline (10s)
resolve_at_msg_count=3, # one more than the published messages count
)
flow_control = types.FlowControl(max_messages=1)
subscription_future = subscriber.subscribe(
subscription_path, callback, flow_control=flow_control
)
# We expect to process the first two messages in 2 * 13 seconds, and
# any duplicate message that is re-sent by the backend in additional
# 13 seconds, totalling 39 seconds (+ overhead) --> if there have been
# no duplicates in 60 seconds, we can reasonably assume that there
# won't be any.
try:
callback.done_future.result(timeout=60)
except exceptions.TimeoutError:
# future timed out, because we received no excessive messages
assert sorted(callback.seen_message_ids) == [1, 2]
else:
> pytest.fail(
"Expected to receive 2 messages, but got at least {}.".format(
len(callback.seen_message_ids)
)
)
E Failed: Expected to receive 2 messages, but got at least 3.
tests/system.py:552: Failed</pre></details>
|
non_process
|
tests system teststreamingpull test streaming pull ack deadline failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output self publisher subscriber project precise truck topic path projects precise truck topics t subscription path projects precise truck subscriptions s cleanup def test streaming pull ack deadline self publisher subscriber project topic path subscription path cleanup make sure the topic and subscription get deleted cleanup append publisher delete topic topic topic path cleanup append subscriber delete subscription subscription subscription path create a topic and a subscription then subscribe to the topic this must happen before the messages are published publisher create topic name topic path subscribe to the topic this must happen before the messages are published subscriber create subscription name subscription path topic topic path ack deadline seconds publish some messages and wait for completion publish messages publisher topic path batch sizes subscribe to the topic callback streamingpullcallback processing time more than the default stream ack deadline resolve at msg count one more than the published messages count flow control types flowcontrol max messages subscription future subscriber subscribe subscription path callback flow control flow control we expect to process the first two messages in seconds and any duplicate message that is re sent by the backend in additional seconds totalling seconds overhead if there have been no duplicates in seconds we can reasonably assume that there won t be any try callback done future result timeout except exceptions timeouterror future timed out because we received no excessive messages assert sorted callback seen message ids else pytest fail expected to receive messages but got at least format len callback seen message ids e failed expected to receive messages but got at least tests system py failed
| 0
|
6,901
| 10,053,667,243
|
IssuesEvent
|
2019-07-21 18:43:56
|
tokio-rs/tokio
|
https://api.github.com/repos/tokio-rs/tokio
|
opened
|
process: Investigate using job objects and IOCP on Windows
|
tokio-process
|
Originally reported in alexcrichton/tokio-process#11:
> While reading the MSDN docs on job objects for something, I came across this page:
https://msdn.microsoft.com/en-us/library/windows/desktop/ms684141(v=vs.85).aspx
>
> If you create a job object, you can associate an IOCP with it. If you then create processes in that job (generally using [`AssignProcessToJobObject`](https://msdn.microsoft.com/en-us/library/windows/desktop/ms681949(v=vs.85).aspx) on a process that has been created with the `CREATE_SUSPENDED` flag) you will get process event messages on the IOCP, including `JOB_OBJECT_MSG_EXIT_PROCESS`.
Please see the comments in the original discussion for context. It appears that this depends on tokio-rs/mio#526
|
1.0
|
process: Investigate using job objects and IOCP on Windows - Originally reported in alexcrichton/tokio-process#11:
> While reading the MSDN docs on job objects for something, I came across this page:
https://msdn.microsoft.com/en-us/library/windows/desktop/ms684141(v=vs.85).aspx
>
> If you create a job object, you can associate an IOCP with it. If you then create processes in that job (generally using [`AssignProcessToJobObject`](https://msdn.microsoft.com/en-us/library/windows/desktop/ms681949(v=vs.85).aspx) on a process that has been created with the `CREATE_SUSPENDED` flag) you will get process event messages on the IOCP, including `JOB_OBJECT_MSG_EXIT_PROCESS`.
Please see the comments in the original discussion for context. It appears that this depends on tokio-rs/mio#526
|
process
|
process investigate using job objects and iocp on windows originally reported in alexcrichton tokio process while reading the msdn docs on job objects for something i came across this page if you create a job object you can associate an iocp with it if you then create processes in that job generally using on a process that has been created with the create suspended flag you will get process event messages on the iocp including job object msg exit process please see the comments in the original discussion for context it appears that this depends on tokio rs mio
| 1
|
654,767
| 21,662,157,185
|
IssuesEvent
|
2022-05-06 20:36:28
|
bounswe/bounswe2022group7
|
https://api.github.com/repos/bounswe/bounswe2022group7
|
closed
|
[Database] Implement "ArtItem", "DiscussionPost", and "Comment" models
|
Status: Completed Priority: Medium Difficulty: Medium Type: Implementation
|
[\_\_init.py\_\_](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/__init__.py) and [models.py](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/models.py) files are created with fundamental functions.
Given those, I shall implement `ArtItem`, `DiscussionPost`, and `Comment` models.
- Implement mentioned models
- Push the updates to the [practice-app/database-base](https://github.com/bounswe/bounswe2022group7/tree/practice-app/database-base) branch
**Deadline: 06/05/2022, 14:00**
**Reviewer: @azizamankenova**
|
1.0
|
[Database] Implement "ArtItem", "DiscussionPost", and "Comment" models - [\_\_init.py\_\_](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/__init__.py) and [models.py](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/models.py) files are created with fundamental functions.
Given those, I shall implement `ArtItem`, `DiscussionPost`, and `Comment` models.
- Implement mentioned models
- Push the updates to the [practice-app/database-base](https://github.com/bounswe/bounswe2022group7/tree/practice-app/database-base) branch
**Deadline: 06/05/2022, 14:00**
**Reviewer: @azizamankenova**
|
non_process
|
implement artitem discussionpost and comment models and files are created with fundamental functions given those i shall implement artitem discussionpost and comment models implement mentioned models push the updates to the branch deadline reviewer azizamankenova
| 0
|
443,781
| 12,799,552,447
|
IssuesEvent
|
2020-07-02 15:34:20
|
bitnami-labs/kubewatch
|
https://api.github.com/repos/bitnami-labs/kubewatch
|
opened
|
Support watching arbitrary custom resources
|
enhancement priority-P1
|
Currently kubewatch supports only watching a built-in list of resource kinds.
Nowadays end users consume many custom resources (CRs) and also new k8s versions keep adding new resources and we shouldn't need to update kubewatch every time such a new resource pops up.
|
1.0
|
Support watching arbitrary custom resources - Currently kubewatch supports only watching a built-in list of resource kinds.
Nowadays end users consume many custom resources (CRs) and also new k8s versions keep adding new resources and we shouldn't need to update kubewatch every time such a new resource pops up.
|
non_process
|
support watching arbitrary custom resources currently kubewatch supports only watching a built in list of resource kinds nowadays end users consume many custom resources crs and also new versions keep adding new resources and we shouldn t need to update kubewatch every time such a new resource pops up
| 0
|
16,021
| 20,188,228,740
|
IssuesEvent
|
2022-02-11 01:19:51
|
savitamittalmsft/WAS-SEC-TEST
|
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
|
opened
|
Develop a security training program
|
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Operational Model & DevOps Roles & Responsibilities
|
<a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/">Develop a security training program</a>
<p><b>Why Consider This?</b></p>
Cybersecurity threats are always evolving and therefore those responsible for organizational information security require specialized, continual, and relevant training to ensure staff maintains the level of competency required to protect, detect, and respond.
<p><b>Context</b></p>
<p><span>A security training program is a component of an overall comprehensive cybersecurity program."nbsp; End users are often the target of cyber attacks and therefore need to be equipped with the knowledge and tools to help protect the organization."nbsp; Security program development and maintenance requires on-going investment and support from leadership to make it a critical component of the organization."nbsp; </span></p>
<p><b>Suggested Actions</b></p>
<p><span>See </span><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span> for more information.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span /></p>
|
1.0
|
Develop a security training program - <a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/">Develop a security training program</a>
<p><b>Why Consider This?</b></p>
Cybersecurity threats are always evolving and therefore those responsible for organizational information security require specialized, continual, and relevant training to ensure staff maintains the level of competency required to protect, detect, and respond.
<p><b>Context</b></p>
<p><span>A security training program is a component of an overall comprehensive cybersecurity program."nbsp; End users are often the target of cyber attacks and therefore need to be equipped with the knowledge and tools to help protect the organization."nbsp; Security program development and maintenance requires on-going investment and support from leadership to make it a critical component of the organization."nbsp; </span></p>
<p><b>Suggested Actions</b></p>
<p><span>See </span><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span> for more information.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span /></p>
|
process
|
develop a security training program why consider this cybersecurity threats are always evolving and therefore those responsible for organizational information security require specialized continual and relevant training to ensure staff maintains the level of competency required to protect detect and respond context a security training program is a component of an overall comprehensive cybersecurity program nbsp end users are often the target of cyber attacks and therefore need to be equipped with the knowledge and tools to help protect the organization nbsp security program development and maintenance requires on going investment and support from leadership to make it a critical component of the organization nbsp suggested actions see how microsoft is transforming its approach to security training for more information learn more how microsoft is transforming its approach to security training
| 1
|
143,276
| 19,177,907,851
|
IssuesEvent
|
2021-12-04 00:04:02
|
samq-ghdemo/js-monorepo
|
https://api.github.com/repos/samq-ghdemo/js-monorepo
|
opened
|
CVE-2014-10064 (High) detected in qs-0.6.6.tgz
|
security vulnerability
|
## CVE-2014-10064 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.6.6.tgz</b></p></summary>
<p>querystring parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p>
<p>Path to dependency file: js-monorepo/NodeGoat/package.json</p>
<p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/zaproxy/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- zaproxy-0.2.0.tgz (Root Library)
- request-2.36.0.tgz
- :x: **qs-0.6.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring.
<p>Publish Date: 2018-05-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064>CVE-2014-10064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-10064">https://nvd.nist.gov/vuln/detail/CVE-2014-10064</a></p>
<p>Release Date: 2018-05-31</p>
<p>Fix Resolution: qs - 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 1.0.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2014-10064","vulnerabilityDetails":"The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2014-10064 (High) detected in qs-0.6.6.tgz - ## CVE-2014-10064 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.6.6.tgz</b></p></summary>
<p>querystring parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p>
<p>Path to dependency file: js-monorepo/NodeGoat/package.json</p>
<p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/zaproxy/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- zaproxy-0.2.0.tgz (Root Library)
- request-2.36.0.tgz
- :x: **qs-0.6.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring.
<p>Publish Date: 2018-05-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064>CVE-2014-10064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-10064">https://nvd.nist.gov/vuln/detail/CVE-2014-10064</a></p>
<p>Release Date: 2018-05-31</p>
<p>Fix Resolution: qs - 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 1.0.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2014-10064","vulnerabilityDetails":"The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz querystring parser library home page a href path to dependency file js monorepo nodegoat package json path to vulnerable library js monorepo nodegoat node modules zaproxy node modules qs package json dependency hierarchy zaproxy tgz root library request tgz x qs tgz vulnerable library found in head commit a href found in base branch main vulnerability details the qs module before does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time an attacker could leverage this to cause a temporary denial of service condition for example in a web application other requests would not be processed while this blocking is occurring publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree zaproxy request qs isminimumfixversionavailable true minimumfixversion qs isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the qs module before does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time an attacker could leverage this to cause a temporary denial of service condition for example in a web application other requests would not be processed while this blocking is occurring vulnerabilityurl
| 0
|
17,067
| 22,505,126,751
|
IssuesEvent
|
2022-06-23 14:54:14
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
[API Proposal]: `Process.Architecture` or `Process.Is64Bit`
|
api-suggestion area-System.Diagnostics.Process
|
### Background and motivation
Working a lot with external process memory, I find it a little tedious having to use [`IsWow64Process`](https://docs.microsoft.com/en-us/windows/win32/api/wow64apiset/nf-wow64apiset-iswow64process) each time I want to infer the architecture of a target process. Not to mention, that function's `out Wow64Process` parameter is heaps confusing.
It would be a lot nicer if I could simply check a `Process.Is64Bit` property.
### API Proposal
```cs
namespace System.Diagnostics;
public class Process
{
public bool Is64Bit { get; }
}
```
### API Usage
```csharp
var proc = Process.GetCurrentProcess();
Console.WriteLine(proc.Is64Bit);
```
### Alternative Designs
```cs
namespace System.Diagnostics;
public class Process
{
public enum ProcessArchitecture
{
x64,
x86,
TheFutureIsNow
}
public ProcessArchitecture Architecture { get; }
}
```
### Risks
_No response_
|
1.0
|
[API Proposal]: `Process.Architecture` or `Process.Is64Bit` - ### Background and motivation
Working a lot with external process memory, I find it a little tedious having to use [`IsWow64Process`](https://docs.microsoft.com/en-us/windows/win32/api/wow64apiset/nf-wow64apiset-iswow64process) each time I want to infer the architecture of a target process. Not to mention, that function's `out Wow64Process` parameter is heaps confusing.
It would be a lot nicer if I could simply check a `Process.Is64Bit` property.
### API Proposal
```cs
namespace System.Diagnostics;
public class Process
{
public bool Is64Bit { get; }
}
```
### API Usage
```csharp
var proc = Process.GetCurrentProcess();
Console.WriteLine(proc.Is64Bit);
```
### Alternative Designs
```cs
namespace System.Diagnostics;
public class Process
{
public enum ProcessArchitecture
{
x64,
x86,
TheFutureIsNow
}
public ProcessArchitecture Architecture { get; }
}
```
### Risks
_No response_
|
process
|
process architecture or process background and motivation working a lot with external process memory i find it a little tedious having to use each time i want to infer the architecture of a target process not to mention that function s out parameter is heaps confusing it would be a lot nicer if i could simply check a process property api proposal cs namespace system diagnostics public class process public bool get api usage csharp var proc process getcurrentprocess console writeline proc alternative designs cs namespace system diagnostics public class process public enum processarchitecture thefutureisnow public processarchitecture architecture get risks no response
| 1
|
433,331
| 12,505,688,934
|
IssuesEvent
|
2020-06-02 11:12:10
|
gitcoinco/web
|
https://api.github.com/repos/gitcoinco/web
|
closed
|
Wonky layout in the "x new funded issues" mails
|
bug priority: backlog
|
**Describe the bug**
Layout between the "funder" and the text is all over the place.
**To Reproduce**
Receive e-mail, read e-mail
**Expected behavior**
A clear line between "owner"/"tasker"/"requester" and text
**Screenshots**

**Desktop (please complete the following information):**
- OS: Mac OS X
- Mail Client: Spark
|
1.0
|
Wonky layout in the "x new funded issues" mails - **Describe the bug**
Layout between the "funder" and the text is all over the place.
**To Reproduce**
Receive e-mail, read e-mail
**Expected behavior**
A clear line between "owner"/"tasker"/"requester" and text
**Screenshots**

**Desktop (please complete the following information):**
- OS: Mac OS X
- Mail Client: Spark
|
non_process
|
wonky layout in the x new funded issues mails describe the bug layout between the funder and the text is all over the place to reproduce receive e mail read e mail expected behavior a clear line between owner tasker requester and text screenshots desktop please complete the following information os mac os x mail client spark
| 0
|
21,505
| 29,670,497,061
|
IssuesEvent
|
2023-06-11 10:49:20
|
giorgiopizz/mkShapesRDF
|
https://api.github.com/repos/giorgiopizz/mkShapesRDF
|
closed
|
[PROC] [SHAPE] Prepare Latinos Hands-On tutorial
|
shape-analysis processor
|
The tutorial should be composed of two parts, both working with UL 2018v9:
1. Using processor, define two reasonable workflows for a double muon dataset and DY NLO
2. Using shape analysis create another example folder that works with latinos root files (as it is for 2016Real) that makes shapes for the DY CR in the double muon channel
|
1.0
|
[PROC] [SHAPE] Prepare Latinos Hands-On tutorial - The tutorial should be composed of two parts, both working with UL 2018v9:
1. Using processor, define two reasonable workflows for a double muon dataset and DY NLO
2. Using shape analysis create another example folder that works with latinos root files (as it is for 2016Real) that makes shapes for the DY CR in the double muon channel
|
process
|
prepare latinos hands on tutorial the tutorial should be composed of two parts both working with ul using processor define two reasonable workflows for a double muon dataset and dy nlo using shape analysis create another example folder that works with latinos root files as it is for that makes shapes for the dy cr in the double muon channel
| 1
|
108,761
| 9,331,088,699
|
IssuesEvent
|
2019-03-28 08:56:31
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits Test failure
|
:Search/Search >test-failure v7.1.0 v8.0.0
|
Fails reliably on master and 7.x (possibly other branches)
```
./gradlew :server:unitTest -Dtests.seed=6AD7B08D850920A6 -Dtests.class=org.elasticsearch.action.search.SearchResponseMergerTests -Dtests.method="testMergeSearchHits" -Dtests.security.manager=true -Dtests.locale=ar-SY -Dtests.timezone=SystemV/MST7 -Dcompiler.java=12 -Druntime.java=8
```
Failure:
```
FAILURE 0.67s | SearchResponseMergerTests.testMergeSearchHits <<< FAILURES!
> Throwable #1: java.lang.AssertionError: expected null, but was:<[Lorg.apache.lucene.search.SortField;@6c4ebb78>
> at __randomizedtesting.SeedInfo.seed([6AD7B08D850920A6:3AC351382A6A6789]:0)
> at org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits(SearchResponseMergerTests.java:520)
> at java.lang.Thread.run(Thread.java:748)
```
|
1.0
|
org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits Test failure - Fails reliably on master and 7.x (possibly other branches)
```
./gradlew :server:unitTest -Dtests.seed=6AD7B08D850920A6 -Dtests.class=org.elasticsearch.action.search.SearchResponseMergerTests -Dtests.method="testMergeSearchHits" -Dtests.security.manager=true -Dtests.locale=ar-SY -Dtests.timezone=SystemV/MST7 -Dcompiler.java=12 -Druntime.java=8
```
Failure:
```
FAILURE 0.67s | SearchResponseMergerTests.testMergeSearchHits <<< FAILURES!
> Throwable #1: java.lang.AssertionError: expected null, but was:<[Lorg.apache.lucene.search.SortField;@6c4ebb78>
> at __randomizedtesting.SeedInfo.seed([6AD7B08D850920A6:3AC351382A6A6789]:0)
> at org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits(SearchResponseMergerTests.java:520)
> at java.lang.Thread.run(Thread.java:748)
```
|
non_process
|
org elasticsearch action search searchresponsemergertests testmergesearchhits test failure fails reliably on master and x possibly other branches gradlew server unittest dtests seed dtests class org elasticsearch action search searchresponsemergertests dtests method testmergesearchhits dtests security manager true dtests locale ar sy dtests timezone systemv dcompiler java druntime java failure failure searchresponsemergertests testmergesearchhits failures throwable java lang assertionerror expected null but was at randomizedtesting seedinfo seed at org elasticsearch action search searchresponsemergertests testmergesearchhits searchresponsemergertests java at java lang thread run thread java
| 0
|
12,609
| 15,012,974,343
|
IssuesEvent
|
2021-02-01 03:07:08
|
topcoder-platform/community-app
|
https://api.github.com/repos/topcoder-platform/community-app
|
closed
|
Recommended checkbox functionality
|
FE ShapeupProcess challenge- recommender-tool
|
Recommended checkbox must not be part of the challenge type group.
It must a standalone checkbox
<img width="436" alt="Screenshot 2021-01-27 at 3 37 03 PM" src="https://user-images.githubusercontent.com/58783823/105976274-f757b700-60b5-11eb-89ae-5fcd8b17a4e6.png">
At present checking the recommended checkbox, deselects the challenge type checkboxes. This must not be the case.
|
1.0
|
Recommended checkbox functionality - Recommended checkbox must not be part of the challenge type group.
It must a standalone checkbox
<img width="436" alt="Screenshot 2021-01-27 at 3 37 03 PM" src="https://user-images.githubusercontent.com/58783823/105976274-f757b700-60b5-11eb-89ae-5fcd8b17a4e6.png">
At present checking the recommended checkbox, deselects the challenge type checkboxes. This must not be the case.
|
process
|
recommended checkbox functionality recommended checkbox must not be part of the challenge type group it must a standalone checkbox img width alt screenshot at pm src at present checking the recommended checkbox deselects the challenge type checkboxes this must not be the case
| 1
|
480,008
| 13,821,856,508
|
IssuesEvent
|
2020-10-13 03:31:55
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
No voting for removing someone from office
|
Category: Laws Priority: Medium Type: Feature
|
It seems like the supreme court that can remove a president from office can not do so by a majority vote of the three, but the members are only able to remove him from office by themselves - each one can alone.
|
1.0
|
No voting for removing someone from office - It seems like the supreme court that can remove a president from office can not do so by a majority vote of the three, but the members are only able to remove him from office by themselves - each one can alone.
|
non_process
|
no voting for removing someone from office it seems like the supreme court that can remove a president from office can not do so by a majority vote of the three but the members are only able to remove him from office by themselves each one can alone
| 0
|
5,476
| 27,363,850,307
|
IssuesEvent
|
2023-02-27 17:37:51
|
aws/aws-sam-cli
|
https://api.github.com/repos/aws/aws-sam-cli
|
closed
|
Bug: samconfig.toml values don't override default values in sam template
|
blocked/more-info-needed blocked/close-if-inactive maintainer/need-followup
|
### Description:
Running a .NET Core 3.1 project containing Lambdas, trying to deploy using sam build and sam deploy. Build is successful, but Deploy fails in AWS because parameter overrides specified in samconfig.toml are not used, even though log for both sam build and sam deploy states that they are found correctly.
### Steps to reproduce:
Running ADO pipeline (attached), deploying SAM template (attached) with samconfig.toml.
### Observed result:
The template generated by sam build has original default values specified for parameters, and they are the ones used during deployment, even though log says : for sam build: Loading configuration values from [uat.['build'].parameters] (env.command_name.section) in config file at '/workspace/1/s/Main/Source/Services/samconfig.toml' and similar for sam deploy
### Expected result:
the template to be deployed with parameters specified in [uat.global.parameters] section in samconfig.toml
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
OS: Lambda in AWS
sam --version: Tool 'amazon.lambda.tools' (version '5.6.2')
AWS region: us-west-2
Attached are logs for both Build and Deploy steps as well as the original template, generated one and samconfig.toml (in templatesAndSamCfg.zip)
[Build-Log.txt](https://github.com/aws/aws-sam-cli/files/10406008/Build-Log.txt)
[Deploy-Log.txt](https://github.com/aws/aws-sam-cli/files/10406009/Deploy-Log.txt)
[samconfig.zip](https://github.com/aws/aws-sam-cli/files/10406011/samconfig.zip)
[templatesAndSamCfg.zip](https://github.com/aws/aws-sam-cli/files/10406018/templatesAndSamCfg.zip)
|
True
|
Bug: samconfig.toml values don't override default values in sam template - ### Description:
Running a .NET Core 3.1 project containing Lambdas, trying to deploy using sam build and sam deploy. Build is successful, but Deploy fails in AWS because parameter overrides specified in samconfig.toml are not used, even though log for both sam build and sam deploy states that they are found correctly.
### Steps to reproduce:
Running ADO pipeline (attached), deploying SAM template (attached) with samconfig.toml.
### Observed result:
The template generated by sam build has original default values specified for parameters, and they are the ones used during deployment, even though log says : for sam build: Loading configuration values from [uat.['build'].parameters] (env.command_name.section) in config file at '/workspace/1/s/Main/Source/Services/samconfig.toml' and similar for sam deploy
### Expected result:
the template to be deployed with parameters specified in [uat.global.parameters] section in samconfig.toml
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
OS: Lambda in AWS
sam --version: Tool 'amazon.lambda.tools' (version '5.6.2')
AWS region: us-west-2
Attached are logs for both Build and Deploy steps as well as the original template, generated one and samconfig.toml (in templatesAndSamCfg.zip)
[Build-Log.txt](https://github.com/aws/aws-sam-cli/files/10406008/Build-Log.txt)
[Deploy-Log.txt](https://github.com/aws/aws-sam-cli/files/10406009/Deploy-Log.txt)
[samconfig.zip](https://github.com/aws/aws-sam-cli/files/10406011/samconfig.zip)
[templatesAndSamCfg.zip](https://github.com/aws/aws-sam-cli/files/10406018/templatesAndSamCfg.zip)
|
non_process
|
bug samconfig toml values don t override default values in sam template description running a net core project containing lambdas trying to deploy using sam build and sam deploy build is successful but deploy fails in aws because parameter overrides specified in samconfig toml are not used even though log for both sam build and sam deploy states that they are found correctly steps to reproduce running ado pipeline attached deploying sam template attached with samconfig toml observed result the template generated by sam build has original default values specified for parameters and they are the ones used during deployment even though log says for sam build loading configuration values from parameters env command name section in config file at workspace s main source services samconfig toml and similar for sam deploy expected result the template to be deployed with parameters specified in section in samconfig toml additional environment details ex windows mac amazon linux etc os lambda in aws sam version tool amazon lambda tools version aws region us west attached are logs for both build and deploy steps as well as the original template generated one and samconfig toml in templatesandsamcfg zip
| 0
|
275,299
| 23,904,726,369
|
IssuesEvent
|
2022-09-08 22:44:55
|
handsontable/hyperformula
|
https://api.github.com/repos/handsontable/hyperformula
|
closed
|
Run performance benchmark automatically for each PR
|
Chore Tests Released Performance Impact: Medium
|
### Description
<!--- [mandatory] Describe the actual behavior and expected behavior -->
Run performance benchmark automatically for each PR. Something similar is done in Spreadsheet Viewer repo. It might be helpful
### Links
- https://github.com/handsontable/hyperformula/pull/938#issuecomment-1090052862
- https://handsoncode.slack.com/archives/D02T462B3GS/p1648823588399299 (internal discussion - accessible by @sequba )
|
1.0
|
Run performance benchmark automatically for each PR - ### Description
<!--- [mandatory] Describe the actual behavior and expected behavior -->
Run performance benchmark automatically for each PR. Something similar is done in Spreadsheet Viewer repo. It might be helpful
### Links
- https://github.com/handsontable/hyperformula/pull/938#issuecomment-1090052862
- https://handsoncode.slack.com/archives/D02T462B3GS/p1648823588399299 (internal discussion - accessible by @sequba )
|
non_process
|
run performance benchmark automatically for each pr description run performance benchmark automatically for each pr something similar is done in spreadsheet viewer repo it might be helpful links internal discussion accessible by sequba
| 0
|
169,901
| 13,164,828,075
|
IssuesEvent
|
2020-08-11 04:57:04
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
closed
|
FNX-12780 ⁃ Intermittent UI test failure - defaultDesktopBookmarksFoldersTest
|
eng:ui-test intermittent-test
|
### Firebase Test Run:
https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7976327018966179987/executions/bs.58e6f78a50568c44/testcases/1/test-cases
https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7618900378390394025/executions/bs.2f5d0fab0c8245d1/testcases/1/test-cases
### Stacktrace:
`androidx.test.espresso.base.DefaultFailureHandler$AssertionFailedWithCauseError: 'is displayed on the screen to the user' doesn't match the selected view.
Expected: is displayed on the screen to the user
Got: "MaterialButton{id=2131362047, res-name=bookmark_folders_sign_in, visibility=VISIBLE, width=992, height=132, has-focus=false, has-focusable=true, has-window-focus=true, is-clickable=true, is-enabled=true, is-focused=false, is-focusable=true, is-layout-requested=false, is-selected=false, layout-params=android.widget.LinearLayout$LayoutParams@dc6d387, tag=null, root-is-layout-requested=false, has-input-connection=false, x=44.0, y=1568.0, text=Sign in to see synced bookmarks, input-type=0, ime-target=false, has-links=false, is-checked=false}"
at dalvik.system.VMStack.getThreadStackTrace(Native Method)
at java.lang.Thread.getStackTrace(Thread.java:1538)
at androidx.test.espresso.base.DefaultFailureHandler.getUserFriendlyError(DefaultFailureHandler.java:16)
at androidx.test.espresso.base.DefaultFailureHandler.handle(DefaultFailureHandler.java:36)
at androidx.test.espresso.ViewInteraction.waitForAndHandleInteractionResults(ViewInteraction.java:103)
at androidx.test.espresso.ViewInteraction.check(ViewInteraction.java:31)
at org.mozilla.fenix.ui.robots.BookmarksRobot.verifySignInToSyncButton(BookmarksRobot.kt:116)
at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:85)
at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:35)
at org.mozilla.fenix.ui.robots.ThreeDotMenuMainRobot$Transition.openBookmarks(ThreeDotMenuMainRobot.kt:128)
at org.mozilla.fenix.ui.BookmarksTest.defaultDesktopBookmarksFoldersTest(BookmarksTest.kt:76)`
### Build: 6/18/20, 6/19/20
|
2.0
|
FNX-12780 ⁃ Intermittent UI test failure - defaultDesktopBookmarksFoldersTest - ### Firebase Test Run:
https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7976327018966179987/executions/bs.58e6f78a50568c44/testcases/1/test-cases
https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7618900378390394025/executions/bs.2f5d0fab0c8245d1/testcases/1/test-cases
### Stacktrace:
`androidx.test.espresso.base.DefaultFailureHandler$AssertionFailedWithCauseError: 'is displayed on the screen to the user' doesn't match the selected view.
Expected: is displayed on the screen to the user
Got: "MaterialButton{id=2131362047, res-name=bookmark_folders_sign_in, visibility=VISIBLE, width=992, height=132, has-focus=false, has-focusable=true, has-window-focus=true, is-clickable=true, is-enabled=true, is-focused=false, is-focusable=true, is-layout-requested=false, is-selected=false, layout-params=android.widget.LinearLayout$LayoutParams@dc6d387, tag=null, root-is-layout-requested=false, has-input-connection=false, x=44.0, y=1568.0, text=Sign in to see synced bookmarks, input-type=0, ime-target=false, has-links=false, is-checked=false}"
at dalvik.system.VMStack.getThreadStackTrace(Native Method)
at java.lang.Thread.getStackTrace(Thread.java:1538)
at androidx.test.espresso.base.DefaultFailureHandler.getUserFriendlyError(DefaultFailureHandler.java:16)
at androidx.test.espresso.base.DefaultFailureHandler.handle(DefaultFailureHandler.java:36)
at androidx.test.espresso.ViewInteraction.waitForAndHandleInteractionResults(ViewInteraction.java:103)
at androidx.test.espresso.ViewInteraction.check(ViewInteraction.java:31)
at org.mozilla.fenix.ui.robots.BookmarksRobot.verifySignInToSyncButton(BookmarksRobot.kt:116)
at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:85)
at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:35)
at org.mozilla.fenix.ui.robots.ThreeDotMenuMainRobot$Transition.openBookmarks(ThreeDotMenuMainRobot.kt:128)
at org.mozilla.fenix.ui.BookmarksTest.defaultDesktopBookmarksFoldersTest(BookmarksTest.kt:76)`
### Build: 6/18/20, 6/19/20
|
non_process
|
fnx ⁃ intermittent ui test failure defaultdesktopbookmarksfolderstest firebase test run stacktrace androidx test espresso base defaultfailurehandler assertionfailedwithcauseerror is displayed on the screen to the user doesn t match the selected view expected is displayed on the screen to the user got materialbutton id res name bookmark folders sign in visibility visible width height has focus false has focusable true has window focus true is clickable true is enabled true is focused false is focusable true is layout requested false is selected false layout params android widget linearlayout layoutparams tag null root is layout requested false has input connection false x y text sign in to see synced bookmarks input type ime target false has links false is checked false at dalvik system vmstack getthreadstacktrace native method at java lang thread getstacktrace thread java at androidx test espresso base defaultfailurehandler getuserfriendlyerror defaultfailurehandler java at androidx test espresso base defaultfailurehandler handle defaultfailurehandler java at androidx test espresso viewinteraction waitforandhandleinteractionresults viewinteraction java at androidx test espresso viewinteraction check viewinteraction java at org mozilla fenix ui robots bookmarksrobot verifysignintosyncbutton bookmarksrobot kt at org mozilla fenix ui bookmarkstest defaultdesktopbookmarksfolderstest invoke bookmarkstest kt at org mozilla fenix ui bookmarkstest defaultdesktopbookmarksfolderstest invoke bookmarkstest kt at org mozilla fenix ui robots threedotmenumainrobot transition openbookmarks threedotmenumainrobot kt at org mozilla fenix ui bookmarkstest defaultdesktopbookmarksfolderstest bookmarkstest kt build
| 0
|
12,480
| 14,949,074,403
|
IssuesEvent
|
2021-01-26 10:59:17
|
cmux/koot
|
https://api.github.com/repos/cmux/koot
|
closed
|
新的生命周期 beforeBuild & afterBuild
|
bundling process enhancement
|
```koot.config.js
module.exports = {
// ...
beforeBuild: async (fullConfig) => {},
afterBuild: async (fullConfig) => {},
// ...
};
```
|
1.0
|
新的生命周期 beforeBuild & afterBuild - ```koot.config.js
module.exports = {
// ...
beforeBuild: async (fullConfig) => {},
afterBuild: async (fullConfig) => {},
// ...
};
```
|
process
|
新的生命周期 beforebuild afterbuild koot config js module exports beforebuild async fullconfig afterbuild async fullconfig
| 1
|
14,351
| 17,374,572,807
|
IssuesEvent
|
2021-07-30 18:49:41
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Unlink functionnality actually seem to apply to release pipelines
|
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
|
> The Link and Unlink functionality applies to build pipelines only. It does not apply to release pipelines.
The "Unlink all" button does exist on "Releases" Pipelines, and is referenced on [another doc](https://docs.microsoft.com/en-us/azure/devops/pipelines/library/task-groups?view=azure-devops). I personally don't see it on a "Pipelines" Pipeline, or what we would call "Build" Pipelines. But then, what is refered as Build Pipelines in the docs are actually under "Pipelines" > "Pipelines" in the Azure DevOps menu, so I can't state with certainty what you are referring to when you say "build pipelines" on this page (I find this to be very confusing and I won't miss an occasion to mention it ^^).
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 1ec7e5a9-e22e-a329-14cf-6e21e65cc85f
* Version Independent ID: c96aaf4e-f5ec-c42f-05cb-e2366167fbc6
* Content: [Process parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/parameters?view=azure-devops)
* Content Source: [docs/pipelines/process/parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/parameters.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Unlink functionnality actually seem to apply to release pipelines -
> The Link and Unlink functionality applies to build pipelines only. It does not apply to release pipelines.
The "Unlink all" button does exist on "Releases" Pipelines, and is referenced on [another doc](https://docs.microsoft.com/en-us/azure/devops/pipelines/library/task-groups?view=azure-devops). I personally don't see it on a "Pipelines" Pipeline, or what we would call "Build" Pipelines. But then, what is refered as Build Pipelines in the docs are actually under "Pipelines" > "Pipelines" in the Azure DevOps menu, so I can't state with certainty what you are referring to when you say "build pipelines" on this page (I find this to be very confusing and I won't miss an occasion to mention it ^^).
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 1ec7e5a9-e22e-a329-14cf-6e21e65cc85f
* Version Independent ID: c96aaf4e-f5ec-c42f-05cb-e2366167fbc6
* Content: [Process parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/parameters?view=azure-devops)
* Content Source: [docs/pipelines/process/parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/parameters.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
unlink functionnality actually seem to apply to release pipelines the link and unlink functionality applies to build pipelines only it does not apply to release pipelines the unlink all button does exist on releases pipelines and is referenced on i personally don t see it on a pipelines pipeline or what we would call build pipelines but then what is refered as build pipelines in the docs are actually under pipelines pipelines in the azure devops menu so i can t state with certainty what you are referring to when you say build pipelines on this page i find this to be very confusing and i won t miss an occasion to mention it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
9,667
| 12,675,347,071
|
IssuesEvent
|
2020-06-19 01:23:21
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Still don't know where Run (build) number is set in classic pipelines
|
Pri1 devops-cicd-process/tech devops/prod
|
I came to this page to find out where the "Build Number" option moved since I can no longer find it to edit in our classic pipelines. After reading this page, I still have no idea where the actual field is (despite a note saying that you can leave it blank).
If it's gone, it should be documented here. If it's supposed to be in the new UI but it's not, there's a bug in the UI and an issue should be opened there.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Still don't know where Run (build) number is set in classic pipelines - I came to this page to find out where the "Build Number" option moved since I can no longer find it to edit in our classic pipelines. After reading this page, I still have no idea where the actual field is (despite a note saying that you can leave it blank).
If it's gone, it should be documented here. If it's supposed to be in the new UI but it's not, there's a bug in the UI and an issue should be opened there.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
still don t know where run build number is set in classic pipelines i came to this page to find out where the build number option moved since i can no longer find it to edit in our classic pipelines after reading this page i still have no idea where the actual field is despite a note saying that you can leave it blank if it s gone it should be documented here if it s supposed to be in the new ui but it s not there s a bug in the ui and an issue should be opened there document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
77,781
| 7,603,563,912
|
IssuesEvent
|
2018-04-29 15:48:58
|
OregonStateRocketry/30k2018-CS-Capstone
|
https://api.github.com/repos/OregonStateRocketry/30k2018-CS-Capstone
|
closed
|
Avionics - rocket is using payload's state machine
|
before practice launch #1 required tests
|
Right now mainRocket is using the state machine designed for mainPayload. This will not work because the main rocket does not go through the same phases. Many stages are the same but others are not. It might be confusing to use logic to separate them.
I think we should probably make sure it does what we want it to, then clone this class and customize it for the rocket phases.
|
1.0
|
Avionics - rocket is using payload's state machine - Right now mainRocket is using the state machine designed for mainPayload. This will not work because the main rocket does not go through the same phases. Many stages are the same but others are not. It might be confusing to use logic to separate them.
I think we should probably make sure it does what we want it to, then clone this class and customize it for the rocket phases.
|
non_process
|
avionics rocket is using payload s state machine right now mainrocket is using the state machine designed for mainpayload this will not work because the main rocket does not go through the same phases many stages are the same but others are not it might be confusing to use logic to separate them i think we should probably make sure it does what we want it to then clone this class and customize it for the rocket phases
| 0
|
593,945
| 18,020,551,057
|
IssuesEvent
|
2021-09-16 18:50:16
|
google/shaka-player
|
https://api.github.com/repos/google/shaka-player
|
closed
|
Networking engine does not provide response HTTP status code
|
enhancement contributions welcome Why didn't we catch this sooner? priority:P3
|
**Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?**
Yes
**What version of Shaka Player are you using?**
3.2.0
**Can you reproduce the issue with our latest release version?**
Yes
**Can you reproduce the issue with the latest code from `master`?**
Yes
**Are you using the demo app or your own custom app?**
Irrelevant
**If custom app, can you reproduce the issue using our demo app?**
Irrelevant
**What browser and OS are you using?**
Irrelevant
**For embedded devices (smart TVs, etc.), what model and firmware version are you using?**
Irrelevant
**What are the manifest and license server URIs?**
Irrelevant
**What configuration are you using? What is the output of `player.getConfiguration()`?**
Irrelevant
**What did you do?**
I was trying to implement "**License Wrapping**", based on the official tutorial, as the result I had the following code:
```js
player.getNetworkingEngine().registerResponseFilter(function (type, response) {
if (!type == shaka.net.NetworkingEngine.RequestType.LICENSE) return;
if (response.status === 200) ...One scenario
if (response.status === 201) ...Another scenario
});
```
Unfortunately, my code did not work and it took some time in order to find out the reason. It seems that _Networking Engine_ does not provide **Response Status** which is weird, because it provides almost all the information.
**What did you expect to happen?**
I was expecting that **response** will include **response.status**.
**What actually happened?**
As I mentioned before, my License Wrapping implementation did not work, because I was not able to compare response.status. I don't know the reason why it was excluded from response object.
|
1.0
|
Networking engine does not provide response HTTP status code - **Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?**
Yes
**What version of Shaka Player are you using?**
3.2.0
**Can you reproduce the issue with our latest release version?**
Yes
**Can you reproduce the issue with the latest code from `master`?**
Yes
**Are you using the demo app or your own custom app?**
Irrelevant
**If custom app, can you reproduce the issue using our demo app?**
Irrelevant
**What browser and OS are you using?**
Irrelevant
**For embedded devices (smart TVs, etc.), what model and firmware version are you using?**
Irrelevant
**What are the manifest and license server URIs?**
Irrelevant
**What configuration are you using? What is the output of `player.getConfiguration()`?**
Irrelevant
**What did you do?**
I was trying to implement "**License Wrapping**", based on the official tutorial, as the result I had the following code:
```js
player.getNetworkingEngine().registerResponseFilter(function (type, response) {
if (!type == shaka.net.NetworkingEngine.RequestType.LICENSE) return;
if (response.status === 200) ...One scenario
if (response.status === 201) ...Another scenario
});
```
Unfortunately, my code did not work and it took some time in order to find out the reason. It seems that _Networking Engine_ does not provide **Response Status** which is weird, because it provides almost all the information.
**What did you expect to happen?**
I was expecting that **response** will include **response.status**.
**What actually happened?**
As I mentioned before, my License Wrapping implementation did not work, because I was not able to compare response.status. I don't know the reason why it was excluded from response object.
|
non_process
|
networking engine does not provide response http status code have you read the and checked for duplicate open issues yes what version of shaka player are you using can you reproduce the issue with our latest release version yes can you reproduce the issue with the latest code from master yes are you using the demo app or your own custom app irrelevant if custom app can you reproduce the issue using our demo app irrelevant what browser and os are you using irrelevant for embedded devices smart tvs etc what model and firmware version are you using irrelevant what are the manifest and license server uris irrelevant what configuration are you using what is the output of player getconfiguration irrelevant what did you do i was trying to implement license wrapping based on the official tutorial as the result i had the following code js player getnetworkingengine registerresponsefilter function type response if type shaka net networkingengine requesttype license return if response status one scenario if response status another scenario unfortunately my code did not work and it took some time in order to find out the reason it seems that networking engine does not provide response status which is weird because it provides almost all the information what did you expect to happen i was expecting that response will include response status what actually happened as i mentioned before my license wrapping implementation did not work because i was not able to compare response status i don t know the reason why it was excluded from response object
| 0
|
8,769
| 11,886,378,192
|
IssuesEvent
|
2020-03-27 21:46:34
|
MicrosoftDocs/vsts-docs
|
https://api.github.com/repos/MicrosoftDocs/vsts-docs
|
closed
|
Step condition doesn't work
|
Pri1 devops-cicd-process/tech devops/prod support-request
|
After https://github.com/Pr0methean/BetterRandom/commit/abd1a4353829ee3725527cd2b4734c686b678597, I still don't get any artifacts after the previous step fails. Why not?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3f151218-9a11-0078-e038-f96198a76143
* Version Independent ID: 09c4d032-62f3-d97c-79d7-6fbfd89910e9
* Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml#feedback)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/conditions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Step condition doesn't work - After https://github.com/Pr0methean/BetterRandom/commit/abd1a4353829ee3725527cd2b4734c686b678597, I still don't get any artifacts after the previous step fails. Why not?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3f151218-9a11-0078-e038-f96198a76143
* Version Independent ID: 09c4d032-62f3-d97c-79d7-6fbfd89910e9
* Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml#feedback)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/conditions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
step condition doesn t work after i still don t get any artifacts after the previous step fails why not document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
13,810
| 16,569,500,969
|
IssuesEvent
|
2021-05-30 05:05:02
|
trpo2021/cw-ip-011_keyboardninja
|
https://api.github.com/repos/trpo2021/cw-ip-011_keyboardninja
|
opened
|
merging branches
|
in process
|
need to carefully merge the branches of the test2 graphics and main, along the way, correct the tests and makefile
|
1.0
|
merging branches - need to carefully merge the branches of the test2 graphics and main, along the way, correct the tests and makefile
|
process
|
merging branches need to carefully merge the branches of the graphics and main along the way correct the tests and makefile
| 1
|
160,877
| 25,248,533,241
|
IssuesEvent
|
2022-11-15 13:02:38
|
Sun-Mountain/lettuceMeetApp
|
https://api.github.com/repos/Sun-Mountain/lettuceMeetApp
|
closed
|
Feature - Make nav bar collapsable when viewed on a smaller screen
|
content-design description-needed
|
<!-- e.g.
Title should be describing the story/feature in one sentences:
- As a team user, I want to be able to move a student from one roster to another.
- Create Title Case component and use for all Nav Items, Section Titles, and Table Headers -->
**Description and related issues -**
<!-- Describe the feature here and link or add dependencies to any applicable issues; include background of the problem being solved, policy, regulation, or business justification -->
**Does this need to be broken up into smaller issues?**
**What is the goal or purpose of this ticket?**
**Mockups or Screenshots**
### Acceptance criteria
| Given | When | Then |
| ----- | ---- | ---- |
| | | |
### This task is done when…
This could include:
- [ ] a list of changes to be made
- [ ] all acceptance criteria are met
- [ ] the description of what a user might be able to accomplish once this is done
- [ ] identify 508 considerations and/or recommendations
---
<!-- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. -->
|
1.0
|
Feature - Make nav bar collapsable when viewed on a smaller screen - <!-- e.g.
Title should be describing the story/feature in one sentences:
- As a team user, I want to be able to move a student from one roster to another.
- Create Title Case component and use for all Nav Items, Section Titles, and Table Headers -->
**Description and related issues -**
<!-- Describe the feature here and link or add dependencies to any applicable issues; include background of the problem being solved, policy, regulation, or business justification -->
**Does this need to be broken up into smaller issues?**
**What is the goal or purpose of this ticket?**
**Mockups or Screenshots**
### Acceptance criteria
| Given | When | Then |
| ----- | ---- | ---- |
| | | |
### This task is done when…
This could include:
- [ ] a list of changes to be made
- [ ] all acceptance criteria are met
- [ ] the description of what a user might be able to accomplish once this is done
- [ ] identify 508 considerations and/or recommendations
---
<!-- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. -->
|
non_process
|
feature make nav bar collapsable when viewed on a smaller screen e g title should be describing the story feature in one sentences as a team user i want to be able to move a student from one roster to another create title case component and use for all nav items section titles and table headers description and related issues does this need to be broken up into smaller issues what is the goal or purpose of this ticket mockups or screenshots acceptance criteria given when then this task is done when… this could include a list of changes to be made all acceptance criteria are met the description of what a user might be able to accomplish once this is done identify considerations and or recommendations
| 0
|
303,226
| 26,194,127,000
|
IssuesEvent
|
2023-01-03 11:52:57
|
navikt/tiltaksgjennomforing-prosess
|
https://api.github.com/repos/navikt/tiltaksgjennomforing-prosess
|
closed
|
Bygg av testing-dokumentfordeling
|
deploy dev-fss testing-dokumentfordeling
|
Kommenter med
>/deploy testing-dokumentfordeling
for å deploye til dev-fss.
Commit: 494723237ef34f1de33270cc3c1b78943f51fb48
|
1.0
|
Bygg av testing-dokumentfordeling - Kommenter med
>/deploy testing-dokumentfordeling
for å deploye til dev-fss.
Commit: 494723237ef34f1de33270cc3c1b78943f51fb48
|
non_process
|
bygg av testing dokumentfordeling kommenter med deploy testing dokumentfordeling for å deploye til dev fss commit
| 0
|
16,043
| 20,190,730,255
|
IssuesEvent
|
2022-02-11 05:01:06
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Create runbook step 2 incorrectly indicates to select Powershell instead of PowerShell Workflow
|
automation/svc triaged cxp doc-enhancement process-automation/subsvc Pri2
|
[Enter feedback here]
Step 2 of the creating a Workbook indicates to create a PowerShell workbook and should be a PowerShell Workflow
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3632c749-8963-f5ed-55ec-28af005780bd
* Version Independent ID: 3ec0f957-e320-7ea7-e5f5-07f543f3c31b
* Content: [Tutorial - Create a PowerShell Workflow runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual#feedback)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-textual.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
1.0
|
Create runbook step 2 incorrectly indicates to select Powershell instead of PowerShell Workflow -
[Enter feedback here]
Step 2 of the creating a Workbook indicates to create a PowerShell workbook and should be a PowerShell Workflow
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3632c749-8963-f5ed-55ec-28af005780bd
* Version Independent ID: 3ec0f957-e320-7ea7-e5f5-07f543f3c31b
* Content: [Tutorial - Create a PowerShell Workflow runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual#feedback)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-textual.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
process
|
create runbook step incorrectly indicates to select powershell instead of powershell workflow step of the creating a workbook indicates to create a powershell workbook and should be a powershell workflow document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
| 1
|
12,995
| 15,359,205,502
|
IssuesEvent
|
2021-03-01 15:36:32
|
edwardsmarc/CASFRI
|
https://api.github.com/repos/edwardsmarc/CASFRI
|
opened
|
Create a DropConstraints.sql script
|
blocker enhancement post-translation process translation
|
So constraints can be droped befere retranslating an inventory (should make translation faster) or before translating a new one which does not fullfil all the constraints.
|
1.0
|
Create a DropConstraints.sql script - So constraints can be droped befere retranslating an inventory (should make translation faster) or before translating a new one which does not fullfil all the constraints.
|
process
|
create a dropconstraints sql script so constraints can be droped befere retranslating an inventory should make translation faster or before translating a new one which does not fullfil all the constraints
| 1
|
12,300
| 14,856,300,329
|
IssuesEvent
|
2021-01-18 13:58:37
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Grouping by required fields should be not have nullable output type
|
bug/2-confirmed kind/bug process/candidate team/client topic: groupBy
|
## Problem
```prisma
type User {
id Int @id
name String
}
```
```ts
const user = prisma.user.groupBy({
by: ['name']
})
user[0].name // string | null
```
## Suggested solution
If `name` can't be null in the schema, then the name field in the result set can't be null. It should be:
```ts
user[0].name // string
```
## Original Ticket
#4893
|
1.0
|
Grouping by required fields should be not have nullable output type - ## Problem
```prisma
type User {
id Int @id
name String
}
```
```ts
const user = prisma.user.groupBy({
by: ['name']
})
user[0].name // string | null
```
## Suggested solution
If `name` can't be null in the schema, then the name field in the result set can't be null. It should be:
```ts
user[0].name // string
```
## Original Ticket
#4893
|
process
|
grouping by required fields should be not have nullable output type problem prisma type user id int id name string ts const user prisma user groupby by user name string null suggested solution if name can t be null in the schema then the name field in the result set can t be null it should be ts user name string original ticket
| 1
|
11,950
| 14,713,115,234
|
IssuesEvent
|
2021-01-05 09:53:59
|
nestauk/sg_covid_impact
|
https://api.github.com/repos/nestauk/sg_covid_impact
|
closed
|
Load and process BRES and IDBR data
|
processing
|
We use getters to load and process BRES and IDBR data into division level.
|
1.0
|
Load and process BRES and IDBR data - We use getters to load and process BRES and IDBR data into division level.
|
process
|
load and process bres and idbr data we use getters to load and process bres and idbr data into division level
| 1
|
18,259
| 24,341,438,806
|
IssuesEvent
|
2022-10-01 19:05:32
|
OpenDataScotland/the_od_bods
|
https://api.github.com/repos/OpenDataScotland/the_od_bods
|
closed
|
Tidy filetypes for datasets
|
good first issue data processing back end
|
Filetypes for datasets are getting a bit out of hand. Currently we display them directly as provided by publisher, but inconsistencies are starting to show (e.g. "ZIP" and ".ZIP")
This should be a relatively easy solution to apply and it would be quite similar to how we tidy licensing and category information already (see https://github.com/OpenDataScotland/the_od_bods/blob/main/merge_data.py)
** Note that in https://opendata.scot/datasets.json the filetype is the resource [Name]. In https://raw.githubusercontent.com/OpenDataScotland/the_od_bods/main/data/merged_output.csv it is the [FileType]
See original conversation in slack: https://opendatascotland.slack.com/archives/C02HEHDL8AY/p1655743400444269
|
1.0
|
Tidy filetypes for datasets - Filetypes for datasets are getting a bit out of hand. Currently we display them directly as provided by publisher, but inconsistencies are starting to show (e.g. "ZIP" and ".ZIP")
This should be a relatively easy solution to apply and it would be quite similar to how we tidy licensing and category information already (see https://github.com/OpenDataScotland/the_od_bods/blob/main/merge_data.py)
** Note that in https://opendata.scot/datasets.json the filetype is the resource [Name]. In https://raw.githubusercontent.com/OpenDataScotland/the_od_bods/main/data/merged_output.csv it is the [FileType]
See original conversation in slack: https://opendatascotland.slack.com/archives/C02HEHDL8AY/p1655743400444269
|
process
|
tidy filetypes for datasets filetypes for datasets are getting a bit out of hand currently we display them directly as provided by publisher but inconsistencies are starting to show e g zip and zip this should be a relatively easy solution to apply and it would be quite similar to how we tidy licensing and category information already see note that in the filetype is the resource in it is the see original conversation in slack
| 1
|
15,652
| 19,846,793,876
|
IssuesEvent
|
2022-01-21 07:38:47
|
ooi-data/CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth
|
https://api.github.com/repos/ooi-data/CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:38:46.523700.
## Details
Flow name: `CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:38:46.523700.
## Details
Flow name: `CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host adcp velocity earth task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
147,373
| 19,520,315,274
|
IssuesEvent
|
2021-12-29 17:10:34
|
mregen/UA-.NetStandardLibrary
|
https://api.github.com/repos/mregen/UA-.NetStandardLibrary
|
closed
|
CVE-2019-1302 (High) detected in microsoft.netcore.app.2.1.0.nupkg
|
security vulnerability
|
## CVE-2019-1302 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>microsoft.netcore.app.2.1.0.nupkg</b></p></summary>
<p>A set of .NET API's that are included in the default .NET Core application model.
caa7b7e2bad98e56a...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg</a></p>
<p>Path to dependency file: /Tests/Opc.Ua.Security.Certificates.Tests/Opc.Ua.Security.Certificates.Tests.csproj</p>
<p>Path to vulnerable library: /microsoft.netcore.app/2.1.0/microsoft.netcore.app.2.1.0.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **microsoft.netcore.app.2.1.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mregen/UA-.NetStandardLibrary/commit/cc7c7249fb08f768b869a09371e53abf3b2c2047">cc7c7249fb08f768b869a09371e53abf3b2c2047</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An elevation of privilege vulnerability exists when a ASP.NET Core web application, created using vulnerable project templates, fails to properly sanitize web requests, aka 'ASP.NET Core Elevation Of Privilege Vulnerability'.
<p>Publish Date: 2019-09-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1302>CVE-2019-1302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/aspnet/Announcements/issues/384">https://github.com/aspnet/Announcements/issues/384</a></p>
<p>Release Date: 2019-09-12</p>
<p>Fix Resolution: Microsoft.AspNetCore.SpaServices - 2.2.1,2.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-1302 (High) detected in microsoft.netcore.app.2.1.0.nupkg - ## CVE-2019-1302 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>microsoft.netcore.app.2.1.0.nupkg</b></p></summary>
<p>A set of .NET API's that are included in the default .NET Core application model.
caa7b7e2bad98e56a...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg</a></p>
<p>Path to dependency file: /Tests/Opc.Ua.Security.Certificates.Tests/Opc.Ua.Security.Certificates.Tests.csproj</p>
<p>Path to vulnerable library: /microsoft.netcore.app/2.1.0/microsoft.netcore.app.2.1.0.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **microsoft.netcore.app.2.1.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mregen/UA-.NetStandardLibrary/commit/cc7c7249fb08f768b869a09371e53abf3b2c2047">cc7c7249fb08f768b869a09371e53abf3b2c2047</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An elevation of privilege vulnerability exists when a ASP.NET Core web application, created using vulnerable project templates, fails to properly sanitize web requests, aka 'ASP.NET Core Elevation Of Privilege Vulnerability'.
<p>Publish Date: 2019-09-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1302>CVE-2019-1302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/aspnet/Announcements/issues/384">https://github.com/aspnet/Announcements/issues/384</a></p>
<p>Release Date: 2019-09-12</p>
<p>Fix Resolution: Microsoft.AspNetCore.SpaServices - 2.2.1,2.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in microsoft netcore app nupkg cve high severity vulnerability vulnerable library microsoft netcore app nupkg a set of net api s that are included in the default net core application model library home page a href path to dependency file tests opc ua security certificates tests opc ua security certificates tests csproj path to vulnerable library microsoft netcore app microsoft netcore app nupkg dependency hierarchy x microsoft netcore app nupkg vulnerable library found in head commit a href vulnerability details an elevation of privilege vulnerability exists when a asp net core web application created using vulnerable project templates fails to properly sanitize web requests aka asp net core elevation of privilege vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore spaservices step up your open source security game with whitesource
| 0
|
594,453
| 18,046,020,883
|
IssuesEvent
|
2021-09-18 22:56:59
|
python/mypy
|
https://api.github.com/repos/python/mypy
|
closed
|
Crash in proper_plugin.py when one argument passed to `isinstance`
|
crash priority-2-low
|
**Crash Report**
When using the `misc/proper_plugin.py` plugin, mypy crashes when only one argument is passed to `isinstance`
**Traceback**
```
test.py:1: error: Too few arguments for "isinstance"
test.py:1: error: INTERNAL ERROR -- Please try using mypy master on Github:
https://mypy.readthedocs.io/en/stable/common_issues.html#using-a-development-mypy-build
Please report a bug at https://github.com/python/mypy/issues
version: 0.910
Traceback (most recent call last):
File "mypy/checkexpr.py", line 3905, in accept
File "mypy/checkexpr.py", line 271, in visit_call_expr
File "mypy/checkexpr.py", line 353, in visit_call_expr_inner
File "mypy/checkexpr.py", line 858, in check_call_expr_with_callee_type
File "mypy/checkexpr.py", line 917, in check_call
File "mypy/checkexpr.py", line 1029, in check_callable_call
File "mypy/checkexpr.py", line 729, in apply_function_plugin
File "/mnt/e/Pranav/Github/mypy/misc/proper_plugin.py", line 37, in isinstance_proper_hook
right = get_proper_type(ctx.arg_types[1][0])
IndexError: list index out of range
test.py:1: : note: use --pdb to drop into pdb
```
**To Reproduce**
test.py:
```py
isinstance(5)
```
Run mypy with `mypy test.py --show-traceback` (or just `mypy test.py`).
**Your Environment**
- Mypy version used: 0.910
- Mypy command-line flags: None necessary
- Mypy configuration options from `mypy.ini` (and other config files):
mypy.ini:
```ini
[mypy]
# Change this path if you're not at the root of the mypy repo
plugins = misc/proper_plugin.py
```
- Python version used: 3.8.5
- Operating system and version: Ubuntu 20.04.2 LTS on WSL
|
1.0
|
Crash in proper_plugin.py when one argument passed to `isinstance` - **Crash Report**
When using the `misc/proper_plugin.py` plugin, mypy crashes when only one argument is passed to `isinstance`
**Traceback**
```
test.py:1: error: Too few arguments for "isinstance"
test.py:1: error: INTERNAL ERROR -- Please try using mypy master on Github:
https://mypy.readthedocs.io/en/stable/common_issues.html#using-a-development-mypy-build
Please report a bug at https://github.com/python/mypy/issues
version: 0.910
Traceback (most recent call last):
File "mypy/checkexpr.py", line 3905, in accept
File "mypy/checkexpr.py", line 271, in visit_call_expr
File "mypy/checkexpr.py", line 353, in visit_call_expr_inner
File "mypy/checkexpr.py", line 858, in check_call_expr_with_callee_type
File "mypy/checkexpr.py", line 917, in check_call
File "mypy/checkexpr.py", line 1029, in check_callable_call
File "mypy/checkexpr.py", line 729, in apply_function_plugin
File "/mnt/e/Pranav/Github/mypy/misc/proper_plugin.py", line 37, in isinstance_proper_hook
right = get_proper_type(ctx.arg_types[1][0])
IndexError: list index out of range
test.py:1: : note: use --pdb to drop into pdb
```
**To Reproduce**
test.py:
```py
isinstance(5)
```
Run mypy with `mypy test.py --show-traceback` (or just `mypy test.py`).
**Your Environment**
- Mypy version used: 0.910
- Mypy command-line flags: None necessary
- Mypy configuration options from `mypy.ini` (and other config files):
mypy.ini:
```ini
[mypy]
# Change this path if you're not at the root of the mypy repo
plugins = misc/proper_plugin.py
```
- Python version used: 3.8.5
- Operating system and version: Ubuntu 20.04.2 LTS on WSL
|
non_process
|
crash in proper plugin py when one argument passed to isinstance crash report when using the misc proper plugin py plugin mypy crashes when only one argument is passed to isinstance traceback test py error too few arguments for isinstance test py error internal error please try using mypy master on github please report a bug at version traceback most recent call last file mypy checkexpr py line in accept file mypy checkexpr py line in visit call expr file mypy checkexpr py line in visit call expr inner file mypy checkexpr py line in check call expr with callee type file mypy checkexpr py line in check call file mypy checkexpr py line in check callable call file mypy checkexpr py line in apply function plugin file mnt e pranav github mypy misc proper plugin py line in isinstance proper hook right get proper type ctx arg types indexerror list index out of range test py note use pdb to drop into pdb to reproduce test py py isinstance run mypy with mypy test py show traceback or just mypy test py your environment mypy version used mypy command line flags none necessary mypy configuration options from mypy ini and other config files mypy ini ini change this path if you re not at the root of the mypy repo plugins misc proper plugin py python version used operating system and version ubuntu lts on wsl
| 0
|
8,538
| 11,713,944,699
|
IssuesEvent
|
2020-03-09 11:18:49
|
kazuwjnlab/cvpaper
|
https://api.github.com/repos/kazuwjnlab/cvpaper
|
opened
|
[cvpaper] CVPR2019 #71 Label Efficient Semi-Supervised Learning via Graph Filtering
|
graph graph convolutional neural network graph signal processing semi-supervised learning
|
## \#71 [Label Efficient Semi-Supervised Learning via Graph Filtering](http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Label_Efficient_Semi-Supervised_Learning_via_Graph_Filtering_CVPR_2019_paper.pdf)
Qimai Li, Xiao-Ming Wu, Han Liu, Xiaotong Zhang, Zhichao Guan
### どんな論文か?
グラフベース半教師あり学習は、ラベル間の接続情報を使用できる点で他の半教師あり学習の手法に比べて有利だが、ラベル伝播の手法が古典的なものであったり、NN系の手法を用いるためにラベル付きデータを大量に必要とするためラベル効率が悪い問題があった。この論文では、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングで特徴抽出を行う。さらに、フィルターの強度でラベル効率を操作することも可能となった。
### 新規性
グラフベース半教師あり学習において、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングによって特徴抽出を行っている点
### 結果
半教師あり分類、回帰、ゼロショット画像認識でほぼほぼ SoTA
### その他(なぜ通ったか?等)
graph signal processing の問題に化けさせる論文は読んだことがないので面白い切り口かと。あと、結果がよい。
![thumb]()
graph, graph convolutional neural network, semi-supervised learning, graph signal processing,
|
1.0
|
[cvpaper] CVPR2019 #71 Label Efficient Semi-Supervised Learning via Graph Filtering - ## \#71 [Label Efficient Semi-Supervised Learning via Graph Filtering](http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Label_Efficient_Semi-Supervised_Learning_via_Graph_Filtering_CVPR_2019_paper.pdf)
Qimai Li, Xiao-Ming Wu, Han Liu, Xiaotong Zhang, Zhichao Guan
### どんな論文か?
グラフベース半教師あり学習は、ラベル間の接続情報を使用できる点で他の半教師あり学習の手法に比べて有利だが、ラベル伝播の手法が古典的なものであったり、NN系の手法を用いるためにラベル付きデータを大量に必要とするためラベル効率が悪い問題があった。この論文では、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングで特徴抽出を行う。さらに、フィルターの強度でラベル効率を操作することも可能となった。
### 新規性
グラフベース半教師あり学習において、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングによって特徴抽出を行っている点
### 結果
半教師あり分類、回帰、ゼロショット画像認識でほぼほぼ SoTA
### その他(なぜ通ったか?等)
graph signal processing の問題に化けさせる論文は読んだことがないので面白い切り口かと。あと、結果がよい。
![thumb]()
graph, graph convolutional neural network, semi-supervised learning, graph signal processing,
|
process
|
label efficient semi supervised learning via graph filtering qimai li xiao ming wu han liu xiaotong zhang zhichao guan どんな論文か? グラフベース半教師あり学習は、ラベル間の接続情報を使用できる点で他の半教師あり学習の手法に比べて有利だが、ラベル伝播の手法が古典的なものであったり、nn系の手法を用いるためにラベル付きデータを大量に必要とするためラベル効率が悪い問題があった。この論文では、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングで特徴抽出を行う。さらに、フィルターの強度でラベル効率を操作することも可能となった。 新規性 グラフベース半教師あり学習において、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングによって特徴抽出を行っている点 結果 半教師あり分類、回帰、ゼロショット画像認識でほぼほぼ sota その他(なぜ通ったか?等) graph signal processing の問題に化けさせる論文は読んだことがないので面白い切り口かと。あと、結果がよい。 graph graph convolutional neural network semi supervised learning graph signal processing
| 1
|
7,051
| 10,210,693,024
|
IssuesEvent
|
2019-08-14 15:17:27
|
pelias/pelias
|
https://api.github.com/repos/pelias/pelias
|
closed
|
Support multiple boundary.country parameters in a request
|
enhancement good first issue help wanted processed
|
#### Here's what I did :innocent:
This came in through support (see desk 882). The user is looking for a way to filter by multiple countries with autocomplete, although this applies to other endpoints as well.
Example query: https://mapzen.com/search/explorer/?query=search&text=ymca&boundary.country=GBR%2CIRL&focus.point.lat=-33.856680&focus.point.lon=151.215281
---
#### Here's what I got :scream_cat:
`[ERROR] GBR,IRL is not a valid ISO2/ISO3 country code`
---
#### Here's what I was expecting :sparkles:
The ability to use something like `boundary.country=GBR,IRL` or `boundary.country=GB&boundary.country=DE` with separate query parameters to filter by multiple countries, similar to`layers` and `sources`.
---
#### Here's what I think could be improved :trophy:
@dianashk confirmed that using multiple countries is not currently supported. If the countries are close enough, then you could use `boundary.rect` param, or do some filtering on your own based on the records.
|
1.0
|
Support multiple boundary.country parameters in a request - #### Here's what I did :innocent:
This came in through support (see desk 882). The user is looking for a way to filter by multiple countries with autocomplete, although this applies to other endpoints as well.
Example query: https://mapzen.com/search/explorer/?query=search&text=ymca&boundary.country=GBR%2CIRL&focus.point.lat=-33.856680&focus.point.lon=151.215281
---
#### Here's what I got :scream_cat:
`[ERROR] GBR,IRL is not a valid ISO2/ISO3 country code`
---
#### Here's what I was expecting :sparkles:
The ability to use something like `boundary.country=GBR,IRL` or `boundary.country=GB&boundary.country=DE` with separate query parameters to filter by multiple countries, similar to`layers` and `sources`.
---
#### Here's what I think could be improved :trophy:
@dianashk confirmed that using multiple countries is not currently supported. If the countries are close enough, then you could use `boundary.rect` param, or do some filtering on your own based on the records.
|
process
|
support multiple boundary country parameters in a request here s what i did innocent this came in through support see desk the user is looking for a way to filter by multiple countries with autocomplete although this applies to other endpoints as well example query here s what i got scream cat gbr irl is not a valid country code here s what i was expecting sparkles the ability to use something like boundary country gbr irl or boundary country gb boundary country de with separate query parameters to filter by multiple countries similar to layers and sources here s what i think could be improved trophy dianashk confirmed that using multiple countries is not currently supported if the countries are close enough then you could use boundary rect param or do some filtering on your own based on the records
| 1
|
11,399
| 14,235,002,051
|
IssuesEvent
|
2020-11-18 14:18:44
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
opened
|
Icons for PH calcs
|
Calculator Process Heating
|
Let me know if they still have the white background and I'll send them in slack
Flue Gas

Wall

Hot Air Leak

Fixture

Opening

Cooling

Atmosphere

Charge Materials
To Do
|
1.0
|
Icons for PH calcs - Let me know if they still have the white background and I'll send them in slack
Flue Gas

Wall

Hot Air Leak

Fixture

Opening

Cooling

Atmosphere

Charge Materials
To Do
|
process
|
icons for ph calcs let me know if they still have the white background and i ll send them in slack flue gas wall hot air leak fixture opening cooling atmosphere charge materials to do
| 1
|
536,257
| 15,706,787,326
|
IssuesEvent
|
2021-03-26 17:54:54
|
yalla-coop/chiltern-music-therapy
|
https://api.github.com/repos/yalla-coop/chiltern-music-therapy
|
opened
|
I see a dashboard as a client
|
back-end front-end priority-3
|
__Wireframe link__
https://www.figma.com/file/CcYmhfnXreAPxlfyEmGsAH/Chiltern-Music-Therapy?node-id=469%3A34091
---
### Acceptance Criteria:
_REMEMBER THAT WHOEVER WORKS ON THIS ISSUE MUST TICK OFF ALL THE POINTS IN THIS LIST UNLESS THERE IS CLEAR AGREEMENT IN THE COMMENTS TO SAY OTHERWISE. **DO NOT REVIEW A PR INVOLVING THIS ISSUE UNLESS THIS HAS BEEN DONE**_
- [ ] Set up page in line with wireframes
- [ ] Api call and database query to fetch
- [ ] Initials of the patient (i.e. first character of each first and last name)
- [ ] Name of therapist
- [ ] The latest programme to be created
- [ ] To confirm, view more is a link that goes to the My Programmes Page
|
1.0
|
I see a dashboard as a client - __Wireframe link__
https://www.figma.com/file/CcYmhfnXreAPxlfyEmGsAH/Chiltern-Music-Therapy?node-id=469%3A34091
---
### Acceptance Criteria:
_REMEMBER THAT WHOEVER WORKS ON THIS ISSUE MUST TICK OFF ALL THE POINTS IN THIS LIST UNLESS THERE IS CLEAR AGREEMENT IN THE COMMENTS TO SAY OTHERWISE. **DO NOT REVIEW A PR INVOLVING THIS ISSUE UNLESS THIS HAS BEEN DONE**_
- [ ] Set up page in line with wireframes
- [ ] Api call and database query to fetch
- [ ] Initials of the patient (i.e. first character of each first and last name)
- [ ] Name of therapist
- [ ] The latest programme to be created
- [ ] To confirm, view more is a link that goes to the My Programmes Page
|
non_process
|
i see a dashboard as a client wireframe link acceptance criteria remember that whoever works on this issue must tick off all the points in this list unless there is clear agreement in the comments to say otherwise do not review a pr involving this issue unless this has been done set up page in line with wireframes api call and database query to fetch initials of the patient i e first character of each first and last name name of therapist the latest programme to be created to confirm view more is a link that goes to the my programmes page
| 0
|
195,273
| 22,300,743,141
|
IssuesEvent
|
2022-06-13 08:29:31
|
Trinadh465/gson_parent-2.8.8
|
https://api.github.com/repos/Trinadh465/gson_parent-2.8.8
|
opened
|
CVE-2020-15250 (Medium) detected in junit-4.12.jar
|
security vulnerability
|
## CVE-2020-15250 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junit-4.12.jar</b></p></summary>
<p>JUnit is a unit testing framework for Java, created by Erich Gamma and Kent Beck.</p>
<p>Library home page: <a href="http://junit.org">http://junit.org</a></p>
<p>Path to dependency file: /gson/build.gradle</p>
<p>Path to vulnerable library: /hes/modules-2/files-2.1/junit/junit/4.12/2973d150c0dc1fefe998f834810d68f278ea58ec/junit-4.12.jar</p>
<p>
Dependency Hierarchy:
- :x: **junit-4.12.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/gson_parent-2.8.8/commit/52f0dd349a1379714ba0cc09a4f1315f471c8ca8">52f0dd349a1379714ba0cc09a4f1315f471c8ca8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In JUnit4 from version 4.7 and before 4.13.1, the test rule TemporaryFolder contains a local information disclosure vulnerability. On Unix like systems, the system's temporary directory is shared between all users on that system. Because of this, when files and directories are written into this directory they are, by default, readable by other users on that same system. This vulnerability does not allow other users to overwrite the contents of these directories or files. This is purely an information disclosure vulnerability. This vulnerability impacts you if the JUnit tests write sensitive information, like API keys or passwords, into the temporary folder, and the JUnit tests execute in an environment where the OS has other untrusted users. Because certain JDK file system APIs were only added in JDK 1.7, this this fix is dependent upon the version of the JDK you are using. For Java 1.7 and higher users: this vulnerability is fixed in 4.13.1. For Java 1.6 and lower users: no patch is available, you must use the workaround below. If you are unable to patch, or are stuck running on Java 1.6, specifying the `java.io.tmpdir` system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability. For more information, including an example of vulnerable code, see the referenced GitHub Security Advisory.
<p>Publish Date: 2020-10-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15250>CVE-2020-15250</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp">https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp</a></p>
<p>Release Date: 2020-10-12</p>
<p>Fix Resolution: 4.13.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-15250 (Medium) detected in junit-4.12.jar - ## CVE-2020-15250 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junit-4.12.jar</b></p></summary>
<p>JUnit is a unit testing framework for Java, created by Erich Gamma and Kent Beck.</p>
<p>Library home page: <a href="http://junit.org">http://junit.org</a></p>
<p>Path to dependency file: /gson/build.gradle</p>
<p>Path to vulnerable library: /hes/modules-2/files-2.1/junit/junit/4.12/2973d150c0dc1fefe998f834810d68f278ea58ec/junit-4.12.jar</p>
<p>
Dependency Hierarchy:
- :x: **junit-4.12.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/gson_parent-2.8.8/commit/52f0dd349a1379714ba0cc09a4f1315f471c8ca8">52f0dd349a1379714ba0cc09a4f1315f471c8ca8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In JUnit4 from version 4.7 and before 4.13.1, the test rule TemporaryFolder contains a local information disclosure vulnerability. On Unix like systems, the system's temporary directory is shared between all users on that system. Because of this, when files and directories are written into this directory they are, by default, readable by other users on that same system. This vulnerability does not allow other users to overwrite the contents of these directories or files. This is purely an information disclosure vulnerability. This vulnerability impacts you if the JUnit tests write sensitive information, like API keys or passwords, into the temporary folder, and the JUnit tests execute in an environment where the OS has other untrusted users. Because certain JDK file system APIs were only added in JDK 1.7, this this fix is dependent upon the version of the JDK you are using. For Java 1.7 and higher users: this vulnerability is fixed in 4.13.1. For Java 1.6 and lower users: no patch is available, you must use the workaround below. If you are unable to patch, or are stuck running on Java 1.6, specifying the `java.io.tmpdir` system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability. For more information, including an example of vulnerable code, see the referenced GitHub Security Advisory.
<p>Publish Date: 2020-10-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15250>CVE-2020-15250</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp">https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp</a></p>
<p>Release Date: 2020-10-12</p>
<p>Fix Resolution: 4.13.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in junit jar cve medium severity vulnerability vulnerable library junit jar junit is a unit testing framework for java created by erich gamma and kent beck library home page a href path to dependency file gson build gradle path to vulnerable library hes modules files junit junit junit jar dependency hierarchy x junit jar vulnerable library found in head commit a href found in base branch master vulnerability details in from version and before the test rule temporaryfolder contains a local information disclosure vulnerability on unix like systems the system s temporary directory is shared between all users on that system because of this when files and directories are written into this directory they are by default readable by other users on that same system this vulnerability does not allow other users to overwrite the contents of these directories or files this is purely an information disclosure vulnerability this vulnerability impacts you if the junit tests write sensitive information like api keys or passwords into the temporary folder and the junit tests execute in an environment where the os has other untrusted users because certain jdk file system apis were only added in jdk this this fix is dependent upon the version of the jdk you are using for java and higher users this vulnerability is fixed in for java and lower users no patch is available you must use the workaround below if you are unable to patch or are stuck running on java specifying the java io tmpdir system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability for more information including an example of vulnerable code see the referenced github security advisory publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
4,978
| 7,808,418,722
|
IssuesEvent
|
2018-06-11 20:07:25
|
GoogleCloudPlatform/google-cloud-cpp
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
|
closed
|
Run integration tests against emulator
|
storage testing type: process
|
To get code coverage statistics that make any sense we need to run the integration tests against some kind of emulator. Maybe we can reuse the python stuff we did for httpbin and add other handlers for the emulator. All we need is to keep buckets and objects in memory.
|
1.0
|
Run integration tests against emulator - To get code coverage statistics that make any sense we need to run the integration tests against some kind of emulator. Maybe we can reuse the python stuff we did for httpbin and add other handlers for the emulator. All we need is to keep buckets and objects in memory.
|
process
|
run integration tests against emulator to get code coverage statistics that make any sense we need to run the integration tests against some kind of emulator maybe we can reuse the python stuff we did for httpbin and add other handlers for the emulator all we need is to keep buckets and objects in memory
| 1
|
640,728
| 20,797,615,564
|
IssuesEvent
|
2022-03-17 10:49:26
|
slsdetectorgroup/slsDetectorPackage
|
https://api.github.com/repos/slsdetectorgroup/slsDetectorPackage
|
closed
|
Receiver: disabled ports write files in 10g
|
action - Bug priority - High status - resolved
|
<!-- Preview changes before submitting -->
<!-- Please fill out everything with an *, as this report will be discarded otherwise -->
<!-- This is a comment, the syntax is a bit different from c++ or bash -->
##### *Distribution:
<!-- RHEL7, RHEL6, Fedora, etc -->
##### *Detector type:
<!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard -->
Receiverv
##### *Software Package Version:
<!-- developer, 4.2.0, 4.1.1, etc -->
##### Priority:
<!-- Super Low, Low, Medium, High, Super High -->
High
##### *Describe the bug
<!-- A clear and concise description of what the bug is -->
Disabled ports write files
##### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
##### To Reproduce
<!-- Steps to reproduce the behavior: -->
<!-- 1. Go to '...' -->
<!-- 2. Click on '....' -->
<!-- 3. Scroll down to '....' -->
<!-- 4. See error -->
##### Screenshots
<!-- If applicable, add screenshots to help explain your problem. -->
##### Additional context
<!-- Add any other context about the problem here. -->
|
1.0
|
Receiver: disabled ports write files in 10g - <!-- Preview changes before submitting -->
<!-- Please fill out everything with an *, as this report will be discarded otherwise -->
<!-- This is a comment, the syntax is a bit different from c++ or bash -->
##### *Distribution:
<!-- RHEL7, RHEL6, Fedora, etc -->
##### *Detector type:
<!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard -->
Receiverv
##### *Software Package Version:
<!-- developer, 4.2.0, 4.1.1, etc -->
##### Priority:
<!-- Super Low, Low, Medium, High, Super High -->
High
##### *Describe the bug
<!-- A clear and concise description of what the bug is -->
Disabled ports write files
##### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
##### To Reproduce
<!-- Steps to reproduce the behavior: -->
<!-- 1. Go to '...' -->
<!-- 2. Click on '....' -->
<!-- 3. Scroll down to '....' -->
<!-- 4. See error -->
##### Screenshots
<!-- If applicable, add screenshots to help explain your problem. -->
##### Additional context
<!-- Add any other context about the problem here. -->
|
non_process
|
receiver disabled ports write files in distribution detector type receiverv software package version priority high describe the bug disabled ports write files expected behavior to reproduce screenshots additional context
| 0
|
2,332
| 5,142,636,883
|
IssuesEvent
|
2017-01-12 13:55:36
|
jimbrown75/Permit-Vision-Enhancements
|
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
|
opened
|
Include Template Verifier and Authoriser signatures in permit created from Template
|
bug Medium Priority Should Fix Verified by PTW Process Lead
|
When creating a permit from a template the template verifier and authoriser should be included on the Permit signatures (electronic and printed)
|
1.0
|
Include Template Verifier and Authoriser signatures in permit created from Template - When creating a permit from a template the template verifier and authoriser should be included on the Permit signatures (electronic and printed)
|
process
|
include template verifier and authoriser signatures in permit created from template when creating a permit from a template the template verifier and authoriser should be included on the permit signatures electronic and printed
| 1
|
129,104
| 10,563,213,043
|
IssuesEvent
|
2019-10-04 20:21:24
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Kuberntes 1.15 as officially supported on Rancher 2.2.x
|
[zube]: To Test team/ca
|
Currently k8s 1.15 is experimental on 2.2.x. We have to enable an official support for it, and drop support for k8s 1.12 with that change.
|
1.0
|
Kuberntes 1.15 as officially supported on Rancher 2.2.x - Currently k8s 1.15 is experimental on 2.2.x. We have to enable an official support for it, and drop support for k8s 1.12 with that change.
|
non_process
|
kuberntes as officially supported on rancher x currently is experimental on x we have to enable an official support for it and drop support for with that change
| 0
|
19,155
| 11,156,698,344
|
IssuesEvent
|
2019-12-25 08:37:23
|
kubesphere/kubesphere
|
https://api.github.com/repos/kubesphere/kubesphere
|
closed
|
kubesphere安装配置的外接es,日志系统未开启,安装后部分pod无法启动
|
area/logging area/microservice
|
安装环境:
4核4G * 3
centos7
没有开启日志系统,配置如下:

安装完成后,其他组件正常,现在istio有两个组件异常

报错信息如下:
```
[root@master2 ~]# kubectl logs jaeger-collector-8698b58b55-gh7h9 -n istio-system
2019/12/12 01:16:37 maxprocs: Leaving GOMAXPROCS=4: CPU quota undefined
{"level":"info","ts":1576113397.8572783,"caller":"flags/service.go:115","msg":"Mounting metrics handler on admin server","route":"/metrics"}
{"level":"info","ts":1576113397.8575335,"caller":"flags/admin.go:108","msg":"Mounting health check on admin server","route":"/"}
{"level":"info","ts":1576113397.8576038,"caller":"flags/admin.go:114","msg":"Starting admin HTTP server","http-port":14269}
{"level":"info","ts":1576113397.857621,"caller":"flags/admin.go:100","msg":"Admin server started","http-port":14269,"health-status":"unavailable"}
{"level":"fatal","ts":1576113402.8768172,"caller":"collector/main.go:89","msg":"Failed to init storage factory","error":"failed to create primary Elasticsearch client: health check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host: no Elasticsearch node available","errorVerbose":"no Elasticsearch node available\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.init.ializers\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:88\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:188\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nhealth check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.(*Client).startupHealthcheck\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:1116\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:244\ngithub.com/jaegertracing/jaeger/pkg/es/config.(*Configuration).NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/pkg/es/config/config.go:100\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:80\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nfailed to create primary Elasticsearch client\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:82\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337","stacktrace":"main.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:89\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200"}
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 3m3s default-scheduler Successfully assigned istio-system/jaeger-query-7f9c7c84c-4dkcc to node4
Normal Pulled 2m59s kubelet, node4 Container image "jaegertracing/jaeger-agent:1.13" already present on machine
Normal Created 2m59s kubelet, node4 Created container jaeger-agent
Normal Started 2m58s kubelet, node4 Started container jaeger-agent
Warning Unhealthy 2m57s kubelet, node4 Readiness probe failed: HTTP probe failed with statuscode: 503
Normal Started 2m5s (x4 over 2m59s) kubelet, node4 Started container jaeger-query
Warning BackOff 82s (x7 over 2m47s) kubelet, node4 Back-off restarting failed container
Normal Pulled 69s (x5 over 3m1s) kubelet, node4 Container image "jaegertracing/jaeger-query:1.13" already present on machine
Normal Created 69s (x5 over 3m) kubelet, node4 Created container jaeger-query
```
感觉是es外接配置异常,麻烦看一下配置
|
1.0
|
kubesphere安装配置的外接es,日志系统未开启,安装后部分pod无法启动 - 安装环境:
4核4G * 3
centos7
没有开启日志系统,配置如下:

安装完成后,其他组件正常,现在istio有两个组件异常

报错信息如下:
```
[root@master2 ~]# kubectl logs jaeger-collector-8698b58b55-gh7h9 -n istio-system
2019/12/12 01:16:37 maxprocs: Leaving GOMAXPROCS=4: CPU quota undefined
{"level":"info","ts":1576113397.8572783,"caller":"flags/service.go:115","msg":"Mounting metrics handler on admin server","route":"/metrics"}
{"level":"info","ts":1576113397.8575335,"caller":"flags/admin.go:108","msg":"Mounting health check on admin server","route":"/"}
{"level":"info","ts":1576113397.8576038,"caller":"flags/admin.go:114","msg":"Starting admin HTTP server","http-port":14269}
{"level":"info","ts":1576113397.857621,"caller":"flags/admin.go:100","msg":"Admin server started","http-port":14269,"health-status":"unavailable"}
{"level":"fatal","ts":1576113402.8768172,"caller":"collector/main.go:89","msg":"Failed to init storage factory","error":"failed to create primary Elasticsearch client: health check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host: no Elasticsearch node available","errorVerbose":"no Elasticsearch node available\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.init.ializers\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:88\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:188\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nhealth check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.(*Client).startupHealthcheck\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:1116\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:244\ngithub.com/jaegertracing/jaeger/pkg/es/config.(*Configuration).NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/pkg/es/config/config.go:100\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:80\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nfailed to create primary Elasticsearch client\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:82\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337","stacktrace":"main.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:89\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200"}
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 3m3s default-scheduler Successfully assigned istio-system/jaeger-query-7f9c7c84c-4dkcc to node4
Normal Pulled 2m59s kubelet, node4 Container image "jaegertracing/jaeger-agent:1.13" already present on machine
Normal Created 2m59s kubelet, node4 Created container jaeger-agent
Normal Started 2m58s kubelet, node4 Started container jaeger-agent
Warning Unhealthy 2m57s kubelet, node4 Readiness probe failed: HTTP probe failed with statuscode: 503
Normal Started 2m5s (x4 over 2m59s) kubelet, node4 Started container jaeger-query
Warning BackOff 82s (x7 over 2m47s) kubelet, node4 Back-off restarting failed container
Normal Pulled 69s (x5 over 3m1s) kubelet, node4 Container image "jaegertracing/jaeger-query:1.13" already present on machine
Normal Created 69s (x5 over 3m) kubelet, node4 Created container jaeger-query
```
感觉是es外接配置异常,麻烦看一下配置
|
non_process
|
kubesphere安装配置的外接es,日志系统未开启,安装后部分pod无法启动 安装环境: 没有开启日志系统,配置如下: 安装完成后,其他组件正常,现在istio有两个组件异常 报错信息如下: kubectl logs jaeger collector n istio system maxprocs leaving gomaxprocs cpu quota undefined level info ts caller flags service go msg mounting metrics handler on admin server route metrics level info ts caller flags admin go msg mounting health check on admin server route level info ts caller flags admin go msg starting admin http server http port level info ts caller flags admin go msg admin server started http port health status unavailable level fatal ts caller collector main go msg failed to init storage factory error failed to create primary elasticsearch client health check timeout head dial tcp lookup elasticsearch logging data kubesphere logging system svc cluster local on no such host no elasticsearch node available errorverbose no elasticsearch node available ngithub com jaegertracing jaeger vendor gopkg in olivere elastic init ializers n t home travis gopath src github com jaegertracing jaeger vendor gopkg in olivere elastic client go nruntime main n t home travis gimme versions linux src runtime proc go nruntime goexit n t home travis gimme versions linux src runtime asm s nhealth check timeout head dial tcp lookup elasticsearch logging data kubesphere logging system svc cluster local on no such host ngithub com jaegertracing jaeger vendor gopkg in olivere elastic client startuphealthcheck n t home travis gopath src github com jaegertracing jaeger vendor gopkg in olivere elastic client go ngithub com jaegertracing jaeger vendor gopkg in olivere elastic newclient n t home travis gopath src github com jaegertracing jaeger vendor gopkg in olivere elastic client go ngithub com jaegertracing jaeger pkg es config configuration newclient n t home travis gopath src github com jaegertracing jaeger pkg es config config go ngithub com jaegertracing jaeger plugin storage es factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage es factory go ngithub com jaegertracing jaeger plugin storage factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage factory go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command executec n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go nruntime main n t home travis gimme versions linux src runtime proc go nruntime goexit n t home travis gimme versions linux src runtime asm s nfailed to create primary elasticsearch client ngithub com jaegertracing jaeger plugin storage es factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage es factory go ngithub com jaegertracing jaeger plugin storage factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage factory go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command executec n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go nruntime main n t home travis gimme versions linux src runtime proc go nruntime goexit n t home travis gimme versions linux src runtime asm s stacktrace main main n t home travis gopath src github com jaegertracing jaeger cmd collector main go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command executec n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go nruntime main n t home travis gimme versions linux src runtime proc go events type reason age from message normal scheduled default scheduler successfully assigned istio system jaeger query to normal pulled kubelet container image jaegertracing jaeger agent already present on machine normal created kubelet created container jaeger agent normal started kubelet started container jaeger agent warning unhealthy kubelet readiness probe failed http probe failed with statuscode normal started over kubelet started container jaeger query warning backoff over kubelet back off restarting failed container normal pulled over kubelet container image jaegertracing jaeger query already present on machine normal created over kubelet created container jaeger query 感觉是es外接配置异常,麻烦看一下配置
| 0
|
9,049
| 12,130,108,057
|
IssuesEvent
|
2020-04-23 00:30:41
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt
|
priority: p2 remove-gcp-devrel-py-tools type: process
|
remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt
|
1.0
|
remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt
|
process
|
remove gcp devrel py tools from appengine standard django requirements test txt remove gcp devrel py tools from appengine standard django requirements test txt
| 1
|
20,596
| 27,264,367,261
|
IssuesEvent
|
2023-02-22 16:55:50
|
googleapis/python-bigquery
|
https://api.github.com/repos/googleapis/python-bigquery
|
closed
|
"https://www.googleapis.com/auth/bigquery" scope is unecessary
|
api: bigquery type: process
|
https://github.com/googleapis/python-bigquery/blob/e1aa9218ad22f85c9a6cab8b61d013779376a582/google/cloud/bigquery/client.py#L229
`"https://www.googleapis.com/auth/cloud-platform"` is a superset of `"https://www.googleapis.com/auth/bigquery"`. We should only need 1, and per the rest of cloud we should use `"https://www.googleapis.com/auth/cloud-platform"`
Note: I'm filing this issue due to https://github.com/googleapis/google-auth-library-python/issues/1204. I believe that even after the issue identified there is addressed (default scopes shouldn't cause user/impersonated credentials to downscope), google-auth might be downscoping credentials unnecessarily because it sees the "bigquery" scope and it's not present by default.
|
1.0
|
"https://www.googleapis.com/auth/bigquery" scope is unecessary - https://github.com/googleapis/python-bigquery/blob/e1aa9218ad22f85c9a6cab8b61d013779376a582/google/cloud/bigquery/client.py#L229
`"https://www.googleapis.com/auth/cloud-platform"` is a superset of `"https://www.googleapis.com/auth/bigquery"`. We should only need 1, and per the rest of cloud we should use `"https://www.googleapis.com/auth/cloud-platform"`
Note: I'm filing this issue due to https://github.com/googleapis/google-auth-library-python/issues/1204. I believe that even after the issue identified there is addressed (default scopes shouldn't cause user/impersonated credentials to downscope), google-auth might be downscoping credentials unnecessarily because it sees the "bigquery" scope and it's not present by default.
|
process
|
scope is unecessary is a superset of we should only need and per the rest of cloud we should use note i m filing this issue due to i believe that even after the issue identified there is addressed default scopes shouldn t cause user impersonated credentials to downscope google auth might be downscoping credentials unnecessarily because it sees the bigquery scope and it s not present by default
| 1
|
6,111
| 8,969,727,395
|
IssuesEvent
|
2019-01-29 11:38:28
|
ec-europa/europa-component-library
|
https://api.github.com/repos/ec-europa/europa-component-library
|
opened
|
[RFC] File component - INNO-1320
|
Improvement RFC process: WIP
|
More attributes can be used on the download links in the file component: https://v2--europa-component-library.netlify.com/ec/components/file/code/
- `download` https://www.w3schools.com/tags/att_a_download.asp
- `hreflang` https://www.w3schools.com/tags/att_a_hreflang.asp
- `lang` https://www.w3schools.com/tags/att_global_lang.asp which is not specific to links but should be used here since the label of the link is in another language
|
1.0
|
[RFC] File component - INNO-1320 - More attributes can be used on the download links in the file component: https://v2--europa-component-library.netlify.com/ec/components/file/code/
- `download` https://www.w3schools.com/tags/att_a_download.asp
- `hreflang` https://www.w3schools.com/tags/att_a_hreflang.asp
- `lang` https://www.w3schools.com/tags/att_global_lang.asp which is not specific to links but should be used here since the label of the link is in another language
|
process
|
file component inno more attributes can be used on the download links in the file component download hreflang lang which is not specific to links but should be used here since the label of the link is in another language
| 1
|
12,453
| 14,935,214,128
|
IssuesEvent
|
2021-01-25 11:37:16
|
smertatli/SWE-573
|
https://api.github.com/repos/smertatli/SWE-573
|
closed
|
Requirement Elicitation
|
done in process
|
Meet with the clients to elicit the requirements for the project to be done.
- What is the business problem we will solve by building this software product?
- Who are the people that need this solution?
- Who are the key stakeholders?
- Why does such a problem exist?
- What does this product improve?
- How would you quantify success of this product?
- What is the most important output of this product for you?
|
1.0
|
Requirement Elicitation - Meet with the clients to elicit the requirements for the project to be done.
- What is the business problem we will solve by building this software product?
- Who are the people that need this solution?
- Who are the key stakeholders?
- Why does such a problem exist?
- What does this product improve?
- How would you quantify success of this product?
- What is the most important output of this product for you?
|
process
|
requirement elicitation meet with the clients to elicit the requirements for the project to be done what is the business problem we will solve by building this software product who are the people that need this solution who are the key stakeholders why does such a problem exist what does this product improve how would you quantify success of this product what is the most important output of this product for you
| 1
|
75,681
| 3,470,980,354
|
IssuesEvent
|
2015-12-23 12:25:08
|
USGS-WiM/SiGLDMS
|
https://api.github.com/repos/USGS-WiM/SiGLDMS
|
closed
|
Organization "Add new" button (CSS)
|
Priority: Medium
|
Changed from "Add New" string to "Add New" button. Needs styling.

|
1.0
|
Organization "Add new" button (CSS) - Changed from "Add New" string to "Add New" button. Needs styling.

|
non_process
|
organization add new button css changed from add new string to add new button needs styling
| 0
|
142,019
| 19,012,457,231
|
IssuesEvent
|
2021-11-23 10:48:02
|
Yann-dv/_Pekocko
|
https://api.github.com/repos/Yann-dv/_Pekocko
|
opened
|
CVE-2020-7608 (Medium) detected in yargs-parser-7.0.0.tgz, yargs-parser-10.1.0.tgz
|
security vulnerability
|
## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-7.0.0.tgz</b>, <b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p>
<p>Path to dependency file: _Pekocko/package.json</p>
<p>Path to vulnerable library: _Pekocko/node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- compiler-cli-7.0.0.tgz (Root Library)
- yargs-9.0.1.tgz
- :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p>
<p>Path to dependency file: _Pekocko/package.json</p>
<p>Path to vulnerable library: _Pekocko/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.10.2.tgz (Root Library)
- webpack-dev-server-3.1.8.tgz
- yargs-12.0.2.tgz
- :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Yann-dv/_Pekocko/commit/29a980e4dad903d391a0354b9cb7c71642e2c2fe">29a980e4dad903d391a0354b9cb7c71642e2c2fe</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7608 (Medium) detected in yargs-parser-7.0.0.tgz, yargs-parser-10.1.0.tgz - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-7.0.0.tgz</b>, <b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p>
<p>Path to dependency file: _Pekocko/package.json</p>
<p>Path to vulnerable library: _Pekocko/node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- compiler-cli-7.0.0.tgz (Root Library)
- yargs-9.0.1.tgz
- :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p>
<p>Path to dependency file: _Pekocko/package.json</p>
<p>Path to vulnerable library: _Pekocko/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.10.2.tgz (Root Library)
- webpack-dev-server-3.1.8.tgz
- yargs-12.0.2.tgz
- :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Yann-dv/_Pekocko/commit/29a980e4dad903d391a0354b9cb7c71642e2c2fe">29a980e4dad903d391a0354b9cb7c71642e2c2fe</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in yargs parser tgz yargs parser tgz cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file pekocko package json path to vulnerable library pekocko node modules angular compiler cli node modules yargs parser package json dependency hierarchy compiler cli tgz root library yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file pekocko package json path to vulnerable library pekocko node modules webpack dev server node modules yargs parser package json dependency hierarchy build angular tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch main vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
19,337
| 4,383,043,706
|
IssuesEvent
|
2016-08-07 08:58:24
|
zcash/zcash
|
https://api.github.com/repos/zcash/zcash
|
closed
|
Finish the list of consensus changes for security auditors
|
consensus protocol documentation SECURITY
|
The security auditors need the [list of consensus changes](https://github.com/zcash/zcash/wiki/Security-Auditor-Quick-Start#list-of-consensus-changes) to be finished before the audit can begin. It would also be nice if those bullet points linked to relevant code (in the security-review-frozen version).
|
1.0
|
Finish the list of consensus changes for security auditors - The security auditors need the [list of consensus changes](https://github.com/zcash/zcash/wiki/Security-Auditor-Quick-Start#list-of-consensus-changes) to be finished before the audit can begin. It would also be nice if those bullet points linked to relevant code (in the security-review-frozen version).
|
non_process
|
finish the list of consensus changes for security auditors the security auditors need the to be finished before the audit can begin it would also be nice if those bullet points linked to relevant code in the security review frozen version
| 0
|
2,645
| 5,425,336,484
|
IssuesEvent
|
2017-03-03 05:37:46
|
FujiXeroxNZ-Wellington/Indigo
|
https://api.github.com/repos/FujiXeroxNZ-Wellington/Indigo
|
closed
|
js Auto validate cannot validate elements in List group
|
0-4-Contract Processing 0-Contract Management bug v1.0
|
Error: Angular-auto-validate: invalid bs3 form structure elements must be wrapped by a form-group class
the above error is thrown when validating the form fields using js-autovalidate library
|
1.0
|
js Auto validate cannot validate elements in List group - Error: Angular-auto-validate: invalid bs3 form structure elements must be wrapped by a form-group class
the above error is thrown when validating the form fields using js-autovalidate library
|
process
|
js auto validate cannot validate elements in list group error angular auto validate invalid form structure elements must be wrapped by a form group class the above error is thrown when validating the form fields using js autovalidate library
| 1
|
156,230
| 19,831,499,138
|
IssuesEvent
|
2022-01-20 12:29:48
|
Dima2021/SecurityShepherd
|
https://api.github.com/repos/Dima2021/SecurityShepherd
|
closed
|
CVE-2020-9493 (High) detected in log4j-1.2.7.jar
|
security vulnerability
|
## CVE-2020-9493 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.7.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /tory/log4j/log4j/1.2.7/log4j-1.2.7.jar,/target/owaspSecurityShepherd/WEB-INF/lib/log4j-1.2.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.7.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.
<p>Publish Date: 2021-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493>CVE-2020-9493</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.7","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.7","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-9493","vulnerabilityDetails":"A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-9493 (High) detected in log4j-1.2.7.jar - ## CVE-2020-9493 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.7.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /tory/log4j/log4j/1.2.7/log4j-1.2.7.jar,/target/owaspSecurityShepherd/WEB-INF/lib/log4j-1.2.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.7.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.
<p>Publish Date: 2021-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493>CVE-2020-9493</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.7","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.7","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-9493","vulnerabilityDetails":"A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jar cve high severity vulnerability vulnerable library jar path to dependency file pom xml path to vulnerable library tory jar target owaspsecurityshepherd web inf lib jar dependency hierarchy x jar vulnerable library found in base branch dev vulnerability details a deserialization flaw was found in apache chainsaw versions prior to which could lead to malicious code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a deserialization flaw was found in apache chainsaw versions prior to which could lead to malicious code execution vulnerabilityurl
| 0
|
11,510
| 14,394,750,249
|
IssuesEvent
|
2020-12-03 02:02:31
|
A01551343/4a
|
https://api.github.com/repos/A01551343/4a
|
closed
|
complete_size_estimating_template
|
process-dashboard
|
- completar el formato de estimacion de LOC con los valores reales obtenidos
|
1.0
|
complete_size_estimating_template - - completar el formato de estimacion de LOC con los valores reales obtenidos
|
process
|
complete size estimating template completar el formato de estimacion de loc con los valores reales obtenidos
| 1
|
140,628
| 18,905,973,963
|
IssuesEvent
|
2021-11-16 09:08:01
|
VerdantSparks/vuetify_ts_aspnetcore_starter
|
https://api.github.com/repos/VerdantSparks/vuetify_ts_aspnetcore_starter
|
closed
|
CVE-2018-8292 (High) detected in microsoft.netcore.app.2.0.0.nupkg, system.net.http.4.3.0.nupkg
|
security vulnerability
|
## CVE-2018-8292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>microsoft.netcore.app.2.0.0.nupkg</b>, <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>
<details><summary><b>microsoft.netcore.app.2.0.0.nupkg</b></p></summary>
<p>A set of .NET API's that are included in the default .NET Core application model.
e8b8861ac7faf042c...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg</a></p>
<p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p>
<p>Path to vulnerable library: ft.netcore.app/2.0.0/microsoft.netcore.app.2.0.0.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **microsoft.netcore.app.2.0.0.nupkg** (Vulnerable Library)
</details>
<details><summary><b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p>
<p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- microsoft.aspnetcore.authentication.jwtbearer.2.0.4.nupkg (Root Library)
- microsoft.identitymodel.protocols.openidconnect.2.1.4.nupkg
- microsoft.identitymodel.protocols.2.1.4.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/VerdantSparks/vuetify_ts_aspnetcore_starter/commit/898883a08965de37520ac5536d5fba3acbd14129">898883a08965de37520ac5536d5fba3acbd14129</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dotnet/announcements/issues/88">https://github.com/dotnet/announcements/issues/88</a></p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-8292 (High) detected in microsoft.netcore.app.2.0.0.nupkg, system.net.http.4.3.0.nupkg - ## CVE-2018-8292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>microsoft.netcore.app.2.0.0.nupkg</b>, <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>
<details><summary><b>microsoft.netcore.app.2.0.0.nupkg</b></p></summary>
<p>A set of .NET API's that are included in the default .NET Core application model.
e8b8861ac7faf042c...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg</a></p>
<p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p>
<p>Path to vulnerable library: ft.netcore.app/2.0.0/microsoft.netcore.app.2.0.0.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **microsoft.netcore.app.2.0.0.nupkg** (Vulnerable Library)
</details>
<details><summary><b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p>
<p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- microsoft.aspnetcore.authentication.jwtbearer.2.0.4.nupkg (Root Library)
- microsoft.identitymodel.protocols.openidconnect.2.1.4.nupkg
- microsoft.identitymodel.protocols.2.1.4.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/VerdantSparks/vuetify_ts_aspnetcore_starter/commit/898883a08965de37520ac5536d5fba3acbd14129">898883a08965de37520ac5536d5fba3acbd14129</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dotnet/announcements/issues/88">https://github.com/dotnet/announcements/issues/88</a></p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in microsoft netcore app nupkg system net http nupkg cve high severity vulnerability vulnerable libraries microsoft netcore app nupkg system net http nupkg microsoft netcore app nupkg a set of net api s that are included in the default net core application model library home page a href path to dependency file vuetify ts aspnetcore starter vuetify ts aspnetcore starter csproj path to vulnerable library ft netcore app microsoft netcore app nupkg dependency hierarchy x microsoft netcore app nupkg vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file vuetify ts aspnetcore starter vuetify ts aspnetcore starter csproj path to vulnerable library usr share dotnet sdk nugetfallbackfolder system net http system net http nupkg dependency hierarchy microsoft aspnetcore authentication jwtbearer nupkg root library microsoft identitymodel protocols openidconnect nupkg microsoft identitymodel protocols nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch master vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system net http microsoft powershell commands utility rc step up your open source security game with whitesource
| 0
|
67,236
| 8,114,933,365
|
IssuesEvent
|
2018-08-15 03:36:11
|
the-tale/the-tale
|
https://api.github.com/repos/the-tale/the-tale
|
closed
|
Карты судьбы: поправить карты прямого начисления влияния
|
comp_cards comp_politics cont_game_designe est_simple type_improvement
|
Дополнить описание перечнем модификаторов, которые должны влиять на начисляемое влияние.
Дописать тесты, которые будут это проверять.
Должны ли хоть какие-то модификаторы влиять на эту величину? В текущий момент получается, что есть карты на влияние в задании, на которые действуют все бонусы и штрафы. А есть карты прямого влияния, которые, по идее, должны просто гарантированно давать указанную сумму.
Возможно, надо изменить количество начисляемого влияняи, чтобы компенсировать отсутствие бонусов.
|
1.0
|
Карты судьбы: поправить карты прямого начисления влияния - Дополнить описание перечнем модификаторов, которые должны влиять на начисляемое влияние.
Дописать тесты, которые будут это проверять.
Должны ли хоть какие-то модификаторы влиять на эту величину? В текущий момент получается, что есть карты на влияние в задании, на которые действуют все бонусы и штрафы. А есть карты прямого влияния, которые, по идее, должны просто гарантированно давать указанную сумму.
Возможно, надо изменить количество начисляемого влияняи, чтобы компенсировать отсутствие бонусов.
|
non_process
|
карты судьбы поправить карты прямого начисления влияния дополнить описание перечнем модификаторов которые должны влиять на начисляемое влияние дописать тесты которые будут это проверять должны ли хоть какие то модификаторы влиять на эту величину в текущий момент получается что есть карты на влияние в задании на которые действуют все бонусы и штрафы а есть карты прямого влияния которые по идее должны просто гарантированно давать указанную сумму возможно надо изменить количество начисляемого влияняи чтобы компенсировать отсутствие бонусов
| 0
|
135,257
| 30,274,098,750
|
IssuesEvent
|
2023-07-07 17:55:12
|
WordPress/openverse
|
https://api.github.com/repos/WordPress/openverse
|
closed
|
Use Elasticsearch 8 locally
|
🟧 priority: high 💻 aspect: code 🧰 goal: internal improvement 🧱 stack: api 🔧 tech: elasticsearch
|
## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
We need to make sure Openverse will work with ElasticSearch 8 prior to upgrading our existing cluster.
## Description
<!-- Describe the feature and how it solves the problem. -->
Update the root docker compose to use version `8.8.2` of ElasticSearch, or whatever the latest version is at the time of experimentation.
https://github.com/WordPress/openverse/blob/2f39baa19fecbbf26fe3be913770c26fe9d57794/docker-compose.yml#L184
Additionally we should carefully review the [migration guide](https://www.elastic.co/guide/en/elasticsearch/reference/current/migrating-8.0.html) to make sure nothing we use has been deprecated or altered.
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
<!-- If you would like to work on this, please comment below separately. -->
|
1.0
|
Use Elasticsearch 8 locally - ## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
We need to make sure Openverse will work with ElasticSearch 8 prior to upgrading our existing cluster.
## Description
<!-- Describe the feature and how it solves the problem. -->
Update the root docker compose to use version `8.8.2` of ElasticSearch, or whatever the latest version is at the time of experimentation.
https://github.com/WordPress/openverse/blob/2f39baa19fecbbf26fe3be913770c26fe9d57794/docker-compose.yml#L184
Additionally we should carefully review the [migration guide](https://www.elastic.co/guide/en/elasticsearch/reference/current/migrating-8.0.html) to make sure nothing we use has been deprecated or altered.
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
<!-- If you would like to work on this, please comment below separately. -->
|
non_process
|
use elasticsearch locally problem we need to make sure openverse will work with elasticsearch prior to upgrading our existing cluster description update the root docker compose to use version of elasticsearch or whatever the latest version is at the time of experimentation additionally we should carefully review the to make sure nothing we use has been deprecated or altered alternatives additional context
| 0
|
376,908
| 26,222,514,223
|
IssuesEvent
|
2023-01-04 15:53:27
|
SandraScherer/EntertainmentInfothek
|
https://api.github.com/repos/SandraScherer/EntertainmentInfothek
|
closed
|
Add sound mix information to series
|
documentation enhancement database program
|
- [x] Add table Series_SoundMix to database
- [x] Add/adapt Series class in EntertainmentDB.dll
- [x] Add tests to EntertainmentDB.Tests
- [x] Add/adapt ContentCreator classes in WikiPageCreator
- [x] Add tests to WikiPageCreator.Tests
- [x] Update documentation
- [x] EntertainmentInfothek_Database.vpp
- [x] EntertainmentInfothek_EntertainmentDB.dll.vpp
- [x] EntertainmentInfothek_WikiPageCreator.vpp
- [x] Doxygen
|
1.0
|
Add sound mix information to series - - [x] Add table Series_SoundMix to database
- [x] Add/adapt Series class in EntertainmentDB.dll
- [x] Add tests to EntertainmentDB.Tests
- [x] Add/adapt ContentCreator classes in WikiPageCreator
- [x] Add tests to WikiPageCreator.Tests
- [x] Update documentation
- [x] EntertainmentInfothek_Database.vpp
- [x] EntertainmentInfothek_EntertainmentDB.dll.vpp
- [x] EntertainmentInfothek_WikiPageCreator.vpp
- [x] Doxygen
|
non_process
|
add sound mix information to series add table series soundmix to database add adapt series class in entertainmentdb dll add tests to entertainmentdb tests add adapt contentcreator classes in wikipagecreator add tests to wikipagecreator tests update documentation entertainmentinfothek database vpp entertainmentinfothek entertainmentdb dll vpp entertainmentinfothek wikipagecreator vpp doxygen
| 0
|
15,408
| 19,598,022,471
|
IssuesEvent
|
2022-01-05 20:27:44
|
opensearch-project/data-prepper
|
https://api.github.com/repos/opensearch-project/data-prepper
|
opened
|
Create an Aggregate Processor
|
plugin - processor
|
**Is your feature request related to a problem? Please describe.**
Data Prepper users need a processor that can be utilized for generic stateful aggregation
**Describe the solution you'd like**
A solution has already been proposed in RFC #699
|
1.0
|
Create an Aggregate Processor - **Is your feature request related to a problem? Please describe.**
Data Prepper users need a processor that can be utilized for generic stateful aggregation
**Describe the solution you'd like**
A solution has already been proposed in RFC #699
|
process
|
create an aggregate processor is your feature request related to a problem please describe data prepper users need a processor that can be utilized for generic stateful aggregation describe the solution you d like a solution has already been proposed in rfc
| 1
|
46,008
| 18,924,532,155
|
IssuesEvent
|
2021-11-17 08:02:01
|
Azure/azure-powershell
|
https://api.github.com/repos/Azure/azure-powershell
|
closed
|
Remove-AzSqlDatabaseSecondary Throws Object reference not set to an instance of an object
|
SQL Service Attention question customer-reported needs-author-feedback no-recent-activity
|
<!--
- Make sure you are able to reproduce this issue on the latest released version of Az
- https://www.powershellgallery.com/packages/Az
- Please search the existing issues to see if there has been a similar issue filed
- For issue related to importing a module, please refer to our troubleshooting guide:
- https://github.com/Azure/azure-powershell/blob/master/documentation/troubleshoot-module-load.md
-->
## Description
Cmdlet fails with both parameters and literals.
## Steps to reproduce
```powershell
Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName $svr.ResourceGroupName -PartnerServerName $svr.ServerName -DatabaseName $drdbname -servername $targetSvr -ResourceGroupName $targetRG
```
## Environment data
<!-- Please run $PSVersionTable and paste the output in the below code block
If running the Docker container image, indicate the tag of the image used and the version of Docker engine-->
```
```
## Module versions
<!-- Please run (Get-Module -ListAvailable) and paste the output in the below code block -->
```powershell
This is failing the same way on both Az.SQL 2.12.0 and 2.16.0
PS /Users/joey> $PsVersionTable
Name Value
---- -----
PSVersion 7.1.2
PSEdition Core
GitCommitId 7.1.2
OS Darwin 20.3.0 Darwin Kernel Version 20.3.0: Thu Jan 21 00:07:06 PST 2021; root:xnu-7195.81.3~1/RELEASE_X86_64
Platform Unix
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
and cloud shell
```
## Debug output
<!-- Set $DebugPreference='Continue' before running the repro and paste the resulting debug stream in the below code block
ATTENTION: Be sure to remove any sensitive information that may be in the logs -->
```
Message : Object reference not set to an instance of an object.
StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String
partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : System.NullReferenceException
InvocationInfo : {Remove-AzSqlDatabaseSecondary}
Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' -PartnerServerName 'dcac-demo' -DatabaseName 'adventureworks2017-dr' -servername 'dcac-demo-dr' -ResourceGroupName 'dcac-sqldr' -verbose -debug
Position : At line:1 char:1
+ Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' …
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
HistoryId : 33
HistoryId: 32
Message : Object reference not set to an instance of an object.
StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String
partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : System.NullReferenceException
InvocationInfo : {Remove-AzSqlDatabaseSecondary}
Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug
Position : At line:1 char:1
+ Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - …
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
HistoryId : 32
HistoryId: 30
Message : Object reference not set to an instance of an object.
StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String
partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : System.NullReferenceException
InvocationInfo : {Remove-AzSqlDatabaseSecondary}
Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug
Position : At line:1 char:1
+ Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - …
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
HistoryId : 30
```
## Error output
<!-- Please run Resolve-AzError and paste the output in the below code block
ATTENTION: Be sure to remove any sensitive information that may be in the logs -->
```
Remove-AzSqlDatabaseSecondary: Object reference not set to an instance of an object.
DEBUG: AzureQoSEvent: CommandName - Remove-AzSqlDatabaseSecondary; IsSuccess - False; Duration - 00:00:00.8716089;; Exception - System.NullReferenceException: Object reference not set to an instance of an object.
at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord();
DEBUG: Finish sending metric.
DEBUG: 4:43:03 PM - RemoveAzureSqlDatabaseSecondary end processing.
```
|
1.0
|
Remove-AzSqlDatabaseSecondary Throws Object reference not set to an instance of an object - <!--
- Make sure you are able to reproduce this issue on the latest released version of Az
- https://www.powershellgallery.com/packages/Az
- Please search the existing issues to see if there has been a similar issue filed
- For issue related to importing a module, please refer to our troubleshooting guide:
- https://github.com/Azure/azure-powershell/blob/master/documentation/troubleshoot-module-load.md
-->
## Description
Cmdlet fails with both parameters and literals.
## Steps to reproduce
```powershell
Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName $svr.ResourceGroupName -PartnerServerName $svr.ServerName -DatabaseName $drdbname -servername $targetSvr -ResourceGroupName $targetRG
```
## Environment data
<!-- Please run $PSVersionTable and paste the output in the below code block
If running the Docker container image, indicate the tag of the image used and the version of Docker engine-->
```
```
## Module versions
<!-- Please run (Get-Module -ListAvailable) and paste the output in the below code block -->
```powershell
This is failing the same way on both Az.SQL 2.12.0 and 2.16.0
PS /Users/joey> $PsVersionTable
Name Value
---- -----
PSVersion 7.1.2
PSEdition Core
GitCommitId 7.1.2
OS Darwin 20.3.0 Darwin Kernel Version 20.3.0: Thu Jan 21 00:07:06 PST 2021; root:xnu-7195.81.3~1/RELEASE_X86_64
Platform Unix
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
and cloud shell
```
## Debug output
<!-- Set $DebugPreference='Continue' before running the repro and paste the resulting debug stream in the below code block
ATTENTION: Be sure to remove any sensitive information that may be in the logs -->
```
Message : Object reference not set to an instance of an object.
StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String
partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : System.NullReferenceException
InvocationInfo : {Remove-AzSqlDatabaseSecondary}
Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' -PartnerServerName 'dcac-demo' -DatabaseName 'adventureworks2017-dr' -servername 'dcac-demo-dr' -ResourceGroupName 'dcac-sqldr' -verbose -debug
Position : At line:1 char:1
+ Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' …
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
HistoryId : 33
HistoryId: 32
Message : Object reference not set to an instance of an object.
StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String
partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : System.NullReferenceException
InvocationInfo : {Remove-AzSqlDatabaseSecondary}
Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug
Position : At line:1 char:1
+ Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - …
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
HistoryId : 32
HistoryId: 30
Message : Object reference not set to an instance of an object.
StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String
partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : System.NullReferenceException
InvocationInfo : {Remove-AzSqlDatabaseSecondary}
Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug
Position : At line:1 char:1
+ Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - …
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
HistoryId : 30
```
## Error output
<!-- Please run Resolve-AzError and paste the output in the below code block
ATTENTION: Be sure to remove any sensitive information that may be in the logs -->
```
Remove-AzSqlDatabaseSecondary: Object reference not set to an instance of an object.
DEBUG: AzureQoSEvent: CommandName - Remove-AzSqlDatabaseSecondary; IsSuccess - False; Duration - 00:00:00.8716089;; Exception - System.NullReferenceException: Object reference not set to an instance of an object.
at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName)
at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity)
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0()
at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord();
DEBUG: Finish sending metric.
DEBUG: 4:43:03 PM - RemoveAzureSqlDatabaseSecondary end processing.
```
|
non_process
|
remove azsqldatabasesecondary throws object reference not set to an instance of an object make sure you are able to reproduce this issue on the latest released version of az please search the existing issues to see if there has been a similar issue filed for issue related to importing a module please refer to our troubleshooting guide description cmdlet fails with both parameters and literals steps to reproduce powershell remove azsqldatabasesecondary partnerresourcegroupname svr resourcegroupname partnerservername svr servername databasename drdbname servername targetsvr resourcegroupname targetrg environment data please run psversiontable and paste the output in the below code block if running the docker container image indicate the tag of the image used and the version of docker engine module versions powershell this is failing the same way on both az sql and ps users joey psversiontable name value psversion psedition core gitcommitid os darwin darwin kernel version thu jan pst root xnu release platform unix pscompatibleversions … psremotingprotocolversion serializationversion wsmanstackversion and cloud shell debug output set debugpreference continue before running the repro and paste the resulting debug stream in the below code block attention be sure to remove any sensitive information that may be in the logs message object reference not set to an instance of an object stacktrace at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception system nullreferenceexception invocationinfo remove azsqldatabasesecondary line remove azsqldatabasesecondary partnerresourcegroupname sqldb demo partnerservername dcac demo databasename dr servername dcac demo dr resourcegroupname dcac sqldr verbose debug position at line char remove azsqldatabasesecondary partnerresourcegroupname sqldb demo … historyid historyid message object reference not set to an instance of an object stacktrace at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft windowsazure commands utilities common azurepscmdlet confirmaction string processmessage string target action action at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception system nullreferenceexception invocationinfo remove azsqldatabasesecondary line remove azsqldatabasesecondary partnerresourcegroupname sqldb demo partnerservername dcac demo databasename dr servername dcac demo dr resourcegroupname dcac sqldr verbose debug position at line char remove azsqldatabasesecondary partnerresourcegroupname sqldb demo … historyid historyid message object reference not set to an instance of an object stacktrace at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft windowsazure commands utilities common azurepscmdlet confirmaction string processmessage string target action action at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception system nullreferenceexception invocationinfo remove azsqldatabasesecondary line remove azsqldatabasesecondary partnerresourcegroupname sqldb demo partnerservername dcac demo databasename dr servername dcac demo dr resourcegroupname dcac sqldr verbose debug position at line char remove azsqldatabasesecondary partnerresourcegroupname sqldb demo … historyid error output please run resolve azerror and paste the output in the below code block attention be sure to remove any sensitive information that may be in the logs remove azsqldatabasesecondary object reference not set to an instance of an object debug azureqosevent commandname remove azsqldatabasesecondary issuccess false duration exception system nullreferenceexception object reference not set to an instance of an object at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord debug finish sending metric debug pm removeazuresqldatabasesecondary end processing
| 0
|
9,923
| 12,963,288,340
|
IssuesEvent
|
2020-07-20 18:32:26
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Passing parameters between triggering pipelines
|
Pri2 devops-cicd-process/tech devops/prod support-request
|
How do I pass (or receive) parameters between pipelines? My triggered pipeline has input parameters that need to be set from the 'build' pipeline that triggers it.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @ashkir
* Microsoft Alias: **ashkir**
|
1.0
|
Passing parameters between triggering pipelines - How do I pass (or receive) parameters between pipelines? My triggered pipeline has input parameters that need to be set from the 'build' pipeline that triggers it.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @ashkir
* Microsoft Alias: **ashkir**
|
process
|
passing parameters between triggering pipelines how do i pass or receive parameters between pipelines my triggered pipeline has input parameters that need to be set from the build pipeline that triggers it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login ashkir microsoft alias ashkir
| 1
|
4,410
| 7,299,077,731
|
IssuesEvent
|
2018-02-26 18:58:08
|
UKHomeOffice/dq-aws-transition
|
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
|
closed
|
Finalise Windows Data Ingest Prod server
|
DQ Data Ingest Production S4 Processing
|
Finalise Windows Data Ingest Prod server
- [x] Terraform Deployment
- [x] Post-Deployment Tasks
- [x] SEQ check running
- [x] Data transfer tasks running
- [x] Swapped in vars
## Acceptance criteria
- [x] Data is flowing in from S3
- [x] Data is being processed by SEQ job
- [x] WS is launching jobs
- [x] Data is flowing out to S3
|
1.0
|
Finalise Windows Data Ingest Prod server - Finalise Windows Data Ingest Prod server
- [x] Terraform Deployment
- [x] Post-Deployment Tasks
- [x] SEQ check running
- [x] Data transfer tasks running
- [x] Swapped in vars
## Acceptance criteria
- [x] Data is flowing in from S3
- [x] Data is being processed by SEQ job
- [x] WS is launching jobs
- [x] Data is flowing out to S3
|
process
|
finalise windows data ingest prod server finalise windows data ingest prod server terraform deployment post deployment tasks seq check running data transfer tasks running swapped in vars acceptance criteria data is flowing in from data is being processed by seq job ws is launching jobs data is flowing out to
| 1
|
418,698
| 12,202,203,356
|
IssuesEvent
|
2020-04-30 08:34:35
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
[0.9.0 staging-1528] Work party: strange template for the first party
|
Priority: Medium Status: Fixed
|
1. Make a new server
2. Place a workbench and make 10 hewn log order
3. Place a contract board and start to create workparty
You will see this strange template

|
1.0
|
[0.9.0 staging-1528] Work party: strange template for the first party - 1. Make a new server
2. Place a workbench and make 10 hewn log order
3. Place a contract board and start to create workparty
You will see this strange template

|
non_process
|
work party strange template for the first party make a new server place a workbench and make hewn log order place a contract board and start to create workparty you will see this strange template
| 0
|
55,615
| 13,647,458,165
|
IssuesEvent
|
2020-09-26 03:25:41
|
TerryCavanagh/diceydungeonsbeta
|
https://api.github.com/repos/TerryCavanagh/diceydungeonsbeta
|
closed
|
UI Overlap When Gold Appears in Combat
|
High Priority v0.5: 21st June Build
|
This was fixed [here](https://github.com/TerryCavanagh/diceydungeons.com/issues/755) but is back again in Witch PU.

|
1.0
|
UI Overlap When Gold Appears in Combat - This was fixed [here](https://github.com/TerryCavanagh/diceydungeons.com/issues/755) but is back again in Witch PU.

|
non_process
|
ui overlap when gold appears in combat this was fixed but is back again in witch pu
| 0
|
32,969
| 13,994,311,777
|
IssuesEvent
|
2020-10-28 00:20:57
|
Azure/azure-sdk-for-java
|
https://api.github.com/repos/Azure/azure-sdk-for-java
|
closed
|
[Service Bus] Fix module-info.java
|
Client Service Bus blocking-release
|
`module-info.java` in Service Bus has many `opens` statements exposing multiple packages. These `opens` statements have to be removed before GA.
|
1.0
|
[Service Bus] Fix module-info.java - `module-info.java` in Service Bus has many `opens` statements exposing multiple packages. These `opens` statements have to be removed before GA.
|
non_process
|
fix module info java module info java in service bus has many opens statements exposing multiple packages these opens statements have to be removed before ga
| 0
|
43,343
| 11,628,285,965
|
IssuesEvent
|
2020-02-27 18:02:10
|
google/flogger
|
https://api.github.com/repos/google/flogger
|
closed
|
Flag --incompatible_load_java_rules_from_bzl will break Flogger in Bazel 1.2.1
|
P3 type=defect
|
Incompatible flag --incompatible_load_java_rules_from_bzl will break Flogger once Bazel 1.2.1 is released.
Please see the following CI builds for more information:
* [:ubuntu: 16.04 (OpenJDK 8)](<a href="https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/342#7fdfdbb2-c0ac-49f7-8fb4-4af9267c90e6" target="_blank">:ubuntu: 16.04 (OpenJDK 8)</a>)
Questions? Please file an issue in https://github.com/bazelbuild/continuous-integration
**Important**: Please do NOT modify the issue title since that might break our tools.
|
1.0
|
Flag --incompatible_load_java_rules_from_bzl will break Flogger in Bazel 1.2.1 - Incompatible flag --incompatible_load_java_rules_from_bzl will break Flogger once Bazel 1.2.1 is released.
Please see the following CI builds for more information:
* [:ubuntu: 16.04 (OpenJDK 8)](<a href="https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/342#7fdfdbb2-c0ac-49f7-8fb4-4af9267c90e6" target="_blank">:ubuntu: 16.04 (OpenJDK 8)</a>)
Questions? Please file an issue in https://github.com/bazelbuild/continuous-integration
**Important**: Please do NOT modify the issue title since that might break our tools.
|
non_process
|
flag incompatible load java rules from bzl will break flogger in bazel incompatible flag incompatible load java rules from bzl will break flogger once bazel is released please see the following ci builds for more information ubuntu openjdk questions please file an issue in important please do not modify the issue title since that might break our tools
| 0
|
7,619
| 10,727,398,402
|
IssuesEvent
|
2019-10-28 11:36:08
|
prisma/photonjs
|
https://api.github.com/repos/prisma/photonjs
|
closed
|
Filtering `null` relationships is not possible
|
bug/2-confirmed kind/bug process/candidate
|
## Schema
```prisma
model User {
id Int @id
address Address?
}
model Address {
id Int @id
}
```
## Code
```ts
const usersWithoutAddress = await photon.users.findMany({
where: {
address: null
}
})
```
This generates an incorrect query under the hood with querying `address: {}` instead of `address: null`.
|
1.0
|
Filtering `null` relationships is not possible - ## Schema
```prisma
model User {
id Int @id
address Address?
}
model Address {
id Int @id
}
```
## Code
```ts
const usersWithoutAddress = await photon.users.findMany({
where: {
address: null
}
})
```
This generates an incorrect query under the hood with querying `address: {}` instead of `address: null`.
|
process
|
filtering null relationships is not possible schema prisma model user id int id address address model address id int id code ts const userswithoutaddress await photon users findmany where address null this generates an incorrect query under the hood with querying address instead of address null
| 1
|
17,560
| 23,372,954,891
|
IssuesEvent
|
2022-08-10 21:54:33
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Tip: You can create an empty environment and reference it from deployment jobs
|
devops/prod cba Pri1 devops-cicd-process/tech
|
The document contains this tip:
"Tip: You can create an empty environment and reference it from deployment jobs"
I would like to do that, meaning I would like to run a deployment job that does not actually move any files but just records the deployment.
However, when I try, I get the error:
"No resource were found in the environment with ID 24 matching the specified criteria: ResourceId , ResourceName , ResourceType VirtualMachine, Tags ."
Where can I find more information about using empty environments?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Create target environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Tip: You can create an empty environment and reference it from deployment jobs - The document contains this tip:
"Tip: You can create an empty environment and reference it from deployment jobs"
I would like to do that, meaning I would like to run a deployment job that does not actually move any files but just records the deployment.
However, when I try, I get the error:
"No resource were found in the environment with ID 24 matching the specified criteria: ResourceId , ResourceName , ResourceType VirtualMachine, Tags ."
Where can I find more information about using empty environments?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Create target environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
tip you can create an empty environment and reference it from deployment jobs the document contains this tip tip you can create an empty environment and reference it from deployment jobs i would like to do that meaning i would like to run a deployment job that does not actually move any files but just records the deployment however when i try i get the error no resource were found in the environment with id matching the specified criteria resourceid resourcename resourcetype virtualmachine tags where can i find more information about using empty environments document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
283,148
| 30,889,604,890
|
IssuesEvent
|
2023-08-04 02:58:43
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
reopened
|
CVE-2020-8649 (Medium) detected in linux-stable-rtv4.1.33
|
Mend: dependency security vulnerability
|
## CVE-2020-8649 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
There is a use-after-free vulnerability in the Linux kernel through 5.5.2 in the vgacon_invert_region function in drivers/video/console/vgacon.c.
<p>Publish Date: 2020-02-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8649>CVE-2020-8649</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-8649">https://www.linuxkernelcves.com/cves/CVE-2020-8649</a></p>
<p>Release Date: 2020-02-06</p>
<p>Fix Resolution: v5.6-rc5,v3.16.83,v4.14.173,v4.19.109,v4.4.216,v4.9.216,v5.4.25,v5.5.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8649 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2020-8649 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
There is a use-after-free vulnerability in the Linux kernel through 5.5.2 in the vgacon_invert_region function in drivers/video/console/vgacon.c.
<p>Publish Date: 2020-02-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8649>CVE-2020-8649</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-8649">https://www.linuxkernelcves.com/cves/CVE-2020-8649</a></p>
<p>Release Date: 2020-02-06</p>
<p>Fix Resolution: v5.6-rc5,v3.16.83,v4.14.173,v4.19.109,v4.4.216,v4.9.216,v5.4.25,v5.5.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers video console vgacon c drivers video console vgacon c vulnerability details there is a use after free vulnerability in the linux kernel through in the vgacon invert region function in drivers video console vgacon c publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
280,432
| 21,260,072,498
|
IssuesEvent
|
2022-04-13 02:31:42
|
mikezimm/SecureScript7
|
https://api.github.com/repos/mikezimm/SecureScript7
|
closed
|
Add Debugging code tips in prop pane setup
|
documentation enhancement Complete
|
It's hard to debug. Add a tips with lessons learned in the web part.
|
1.0
|
Add Debugging code tips in prop pane setup - It's hard to debug. Add a tips with lessons learned in the web part.
|
non_process
|
add debugging code tips in prop pane setup it s hard to debug add a tips with lessons learned in the web part
| 0
|
87,780
| 25,209,199,648
|
IssuesEvent
|
2022-11-14 01:02:52
|
OpenBB-finance/OpenBBTerminal
|
https://api.github.com/repos/OpenBB-finance/OpenBBTerminal
|
closed
|
[Bug] Windows Installer Broken with Poetry Install
|
bug build installer
|
**Describe the bug**
Windows installer is broken. [Link](https://github.com/OpenBB-finance/OpenBBTerminal/actions/runs/3286347083/jobs/5414352853)
<img width="1340" alt="Screen Shot 2022-10-20 at 2 10 09 PM" src="https://user-images.githubusercontent.com/53658028/197025636-54437cb9-7ae1-4934-a4e2-96ae52db3a49.png">
|
1.0
|
[Bug] Windows Installer Broken with Poetry Install - **Describe the bug**
Windows installer is broken. [Link](https://github.com/OpenBB-finance/OpenBBTerminal/actions/runs/3286347083/jobs/5414352853)
<img width="1340" alt="Screen Shot 2022-10-20 at 2 10 09 PM" src="https://user-images.githubusercontent.com/53658028/197025636-54437cb9-7ae1-4934-a4e2-96ae52db3a49.png">
|
non_process
|
windows installer broken with poetry install describe the bug windows installer is broken img width alt screen shot at pm src
| 0
|
10,640
| 13,446,159,186
|
IssuesEvent
|
2020-09-08 12:33:48
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
PARS - Design review for search results display
|
EPIC - PARs process NEW :new:
|
TASK
Design tweaks needed to the search results display, so that they are in line with SPC / PILs
### Acceptance Criteria
- [ ] The commas are removed from the between product name, dose, strength etc
- [ ] There is a dash between the name and the PL number (example...... Ibuprofen 200mg Tablets (Ibuprofen) - PL 17907/0160)
**Customer acceptance criteria**
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
1.0
|
PARS - Design review for search results display - TASK
Design tweaks needed to the search results display, so that they are in line with SPC / PILs
### Acceptance Criteria
- [ ] The commas are removed from the between product name, dose, strength etc
- [ ] There is a dash between the name and the PL number (example...... Ibuprofen 200mg Tablets (Ibuprofen) - PL 17907/0160)
**Customer acceptance criteria**
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
process
|
pars design review for search results display task design tweaks needed to the search results display so that they are in line with spc pils acceptance criteria the commas are removed from the between product name dose strength etc there is a dash between the name and the pl number example ibuprofen tablets ibuprofen pl customer acceptance criteria technical acceptance criteria data acceptance criteria testing acceptance criteria size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
| 1
|
2,635
| 5,412,522,136
|
IssuesEvent
|
2017-03-01 14:45:38
|
DynareTeam/dynare
|
https://api.github.com/repos/DynareTeam/dynare
|
closed
|
create fortran preprocessor output
|
enhancement preprocessor
|
dynamic and static files each as independent functions
model as a fortran module
|
1.0
|
create fortran preprocessor output - dynamic and static files each as independent functions
model as a fortran module
|
process
|
create fortran preprocessor output dynamic and static files each as independent functions model as a fortran module
| 1
|
93,849
| 19,346,556,203
|
IssuesEvent
|
2021-12-15 11:25:55
|
Onelinerhub/onelinerhub
|
https://api.github.com/repos/Onelinerhub/onelinerhub
|
closed
|
Short solution needed: "How to list docker tags" (docker)
|
help wanted good first issue code docker
|
Please help us write most modern and shortest code solution for this issue:
**How to list docker tags** (technology: [docker](https://onelinerhub.com/docker))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to use comments to make solution explained.
3. Link to this issue in comments of pull request.
|
1.0
|
Short solution needed: "How to list docker tags" (docker) - Please help us write most modern and shortest code solution for this issue:
**How to list docker tags** (technology: [docker](https://onelinerhub.com/docker))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to use comments to make solution explained.
3. Link to this issue in comments of pull request.
|
non_process
|
short solution needed how to list docker tags docker please help us write most modern and shortest code solution for this issue how to list docker tags technology fast way just write the code solution in the comments prefered way create pull request with a new code file inside don t forget to use comments to make solution explained link to this issue in comments of pull request
| 0
|
249,185
| 7,953,975,332
|
IssuesEvent
|
2018-07-12 05:11:51
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
Placed items disappear but block placement
|
Medium Priority
|
**Version:** 0.7.4.5 beta
**Expected behavior:** When placing an item (so far tried ramp, chair), it should appear.
**Actual behavior:** The item disappears immediately. I can walk through it, but I cant place anything else there (blocks or items). The item can be deleted with the Dev Tool (ramps need to be deleted block by block)
**Steps to reproduce:** In the savegame, in the area around x=373 y=210 (bothin a tunnel at z=47 and on a mountain at z=56) try placing an item (tested with ramps, chairs), item disappears, placement of items or blocks not possible
Invisible item can be deleted with the DevTool, but the same problem occurs again when placing a new item




|
1.0
|
Placed items disappear but block placement - **Version:** 0.7.4.5 beta
**Expected behavior:** When placing an item (so far tried ramp, chair), it should appear.
**Actual behavior:** The item disappears immediately. I can walk through it, but I cant place anything else there (blocks or items). The item can be deleted with the Dev Tool (ramps need to be deleted block by block)
**Steps to reproduce:** In the savegame, in the area around x=373 y=210 (bothin a tunnel at z=47 and on a mountain at z=56) try placing an item (tested with ramps, chairs), item disappears, placement of items or blocks not possible
Invisible item can be deleted with the DevTool, but the same problem occurs again when placing a new item




|
non_process
|
placed items disappear but block placement version beta expected behavior when placing an item so far tried ramp chair it should appear actual behavior the item disappears immediately i can walk through it but i cant place anything else there blocks or items the item can be deleted with the dev tool ramps need to be deleted block by block steps to reproduce in the savegame in the area around x y bothin a tunnel at z and on a mountain at z try placing an item tested with ramps chairs item disappears placement of items or blocks not possible invisible item can be deleted with the devtool but the same problem occurs again when placing a new item
| 0
|
20,950
| 27,809,757,770
|
IssuesEvent
|
2023-03-18 01:38:59
|
serai-dex/serai
|
https://api.github.com/repos/serai-dex/serai
|
closed
|
develop CI is failing
|
bug CI/CD processor bitcoin
|
It looks like the Bitcoin node the processor expects isn't being reached/isn't configured as expected?
|
1.0
|
develop CI is failing - It looks like the Bitcoin node the processor expects isn't being reached/isn't configured as expected?
|
process
|
develop ci is failing it looks like the bitcoin node the processor expects isn t being reached isn t configured as expected
| 1
|
9,184
| 12,228,607,382
|
IssuesEvent
|
2020-05-03 20:09:42
|
chfor183/data_science_articles
|
https://api.github.com/repos/chfor183/data_science_articles
|
opened
|
Imbalanced Datasets
|
Data Preprocessing
|
## TL;DR
Yes !
## Key Takeaways
- 1
- 2
## Useful Code Snippets
```
function test() {
console.log("notice the blank line before this function?");
}
```
## Articles/Ressources
https://towardsdatascience.com/handling-imbalanced-datasets-in-machine-learning-7a0e84220f28
https://towardsdatascience.com/imbalanced-class-sizes-and-classification-models-a-cautionary-tale-3648b8586e03
|
1.0
|
Imbalanced Datasets - ## TL;DR
Yes !
## Key Takeaways
- 1
- 2
## Useful Code Snippets
```
function test() {
console.log("notice the blank line before this function?");
}
```
## Articles/Ressources
https://towardsdatascience.com/handling-imbalanced-datasets-in-machine-learning-7a0e84220f28
https://towardsdatascience.com/imbalanced-class-sizes-and-classification-models-a-cautionary-tale-3648b8586e03
|
process
|
imbalanced datasets tl dr yes key takeaways useful code snippets function test console log notice the blank line before this function articles ressources
| 1
|
35,818
| 6,499,122,969
|
IssuesEvent
|
2017-08-22 20:12:15
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Please note versions in Bazel docs
|
P2 type: documentation
|
On TensorFlow and Closure Rules, I want to be able to provide at least some level of support for old versions of Bazel.
This is currently not easy to do, because the documentation doesn't tell me at which version a feature was introduced, or became deprecated.
For example, [ConfigurationTransition](https://www.bazel.io/versions/master/docs/skylark/lib/ConfigurationTransition.html) doesn't say at which version `"host"` instead of `HOST_CFG` started working.
Or the documentation for [alias()](https://www.bazel.io/versions/master/docs/be/general.html#alias) doesn't tell me in which version it was introduced.
I understand Bazel isn't quite at 1.0 yet. But I still think taking the extra time to do this would be worthwhile.
|
1.0
|
Please note versions in Bazel docs - On TensorFlow and Closure Rules, I want to be able to provide at least some level of support for old versions of Bazel.
This is currently not easy to do, because the documentation doesn't tell me at which version a feature was introduced, or became deprecated.
For example, [ConfigurationTransition](https://www.bazel.io/versions/master/docs/skylark/lib/ConfigurationTransition.html) doesn't say at which version `"host"` instead of `HOST_CFG` started working.
Or the documentation for [alias()](https://www.bazel.io/versions/master/docs/be/general.html#alias) doesn't tell me in which version it was introduced.
I understand Bazel isn't quite at 1.0 yet. But I still think taking the extra time to do this would be worthwhile.
|
non_process
|
please note versions in bazel docs on tensorflow and closure rules i want to be able to provide at least some level of support for old versions of bazel this is currently not easy to do because the documentation doesn t tell me at which version a feature was introduced or became deprecated for example doesn t say at which version host instead of host cfg started working or the documentation for doesn t tell me in which version it was introduced i understand bazel isn t quite at yet but i still think taking the extra time to do this would be worthwhile
| 0
|
205,977
| 15,702,893,815
|
IssuesEvent
|
2021-03-26 13:13:24
|
OpenLiberty/open-liberty
|
https://api.github.com/repos/OpenLiberty/open-liberty
|
opened
|
Feature Test Summary
|
Feature Test Summary team:Zombie Apocalypse
|
## Test Strategy
Test strategy followed what was done for the addition of a very similar property that allowed the sharing of connections on direct lookups.
### Test strategy
* What functionality is new or modified by this feature? Feature adds a boolean attribute `enableContainerAuthForDirectLookups` to existing connectionManager config element. If the value of the new attrib is true, when a datasource is looked up w/o injection or resource-ref, the default authentication mechanism will be CONTAINER instead of APPLICATION. Prior to this feature, the only way to change the autheniticaton was via a resource -ref.
* What are the positive and negative tests for that functionality? Since the default behavior is unchanged, ie, APPLICATION auth, and we have tests that perform direct lookups as part of other tests, the negative test path is already covered. For the positive, a test was added to the com.ibm.ws.jdbc_fat test suite that sets the new attrib to true, performs a lookup, obtains a connection and examines the credentials used by the database to make certain the specified container auth user was used to login and not the default.
* What manual tests are there (if any)? None.
## Confidence Level
**Collectively as a team you need to assess your confidence in the testing delivered based on the values below. This should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole.**
Please indicate your confidence in the testing (up to and including FAT) delivered with this feature by selecting one of these values:
0 - No automated testing delivered
1 - We have minimal automated coverage of the feature including golden paths. There is a relatively high risk that defects or issues could be found in this feature.
2 - We have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here. Error/outlying scenarios are not really covered. There are likely risks that issues may exist in the golden paths
3 - We have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error/outlying scenarios. There is a risk when the feature is used outside the golden paths however we are confident on the golden path. Note: This may still be a valid end state for a feature... things like Beta features may well suffice at this level.
4 - We have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error/outlying scenarios. While more testing of the error/outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide.
5 - We have delivered all automated testing we believe is needed for this feature. The testing covers all golden path cases as well as all the error/outlying scenarios that make sense. We are not aware of any gaps in the testing at this time. No manual testing is required to verify this feature.
Based on your answer above, for any answer other than a 4 or 5 please provide details of what drove your answer. Please be aware, it may be perfectly reasonable in some scenarios to deliver with any value above. We may accept no automated testing is needed for some features, we may be happy with low levels of testing on samples for instance so please don't feel the need to drive to a 5. We need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid. What are the gaps, what is the risk etc. Please also provide links to the follow on work that is needed to close the gaps (should you deem it needed)
|
1.0
|
Feature Test Summary - ## Test Strategy
Test strategy followed what was done for the addition of a very similar property that allowed the sharing of connections on direct lookups.
### Test strategy
* What functionality is new or modified by this feature? Feature adds a boolean attribute `enableContainerAuthForDirectLookups` to existing connectionManager config element. If the value of the new attrib is true, when a datasource is looked up w/o injection or resource-ref, the default authentication mechanism will be CONTAINER instead of APPLICATION. Prior to this feature, the only way to change the autheniticaton was via a resource -ref.
* What are the positive and negative tests for that functionality? Since the default behavior is unchanged, ie, APPLICATION auth, and we have tests that perform direct lookups as part of other tests, the negative test path is already covered. For the positive, a test was added to the com.ibm.ws.jdbc_fat test suite that sets the new attrib to true, performs a lookup, obtains a connection and examines the credentials used by the database to make certain the specified container auth user was used to login and not the default.
* What manual tests are there (if any)? None.
## Confidence Level
**Collectively as a team you need to assess your confidence in the testing delivered based on the values below. This should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole.**
Please indicate your confidence in the testing (up to and including FAT) delivered with this feature by selecting one of these values:
0 - No automated testing delivered
1 - We have minimal automated coverage of the feature including golden paths. There is a relatively high risk that defects or issues could be found in this feature.
2 - We have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here. Error/outlying scenarios are not really covered. There are likely risks that issues may exist in the golden paths
3 - We have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error/outlying scenarios. There is a risk when the feature is used outside the golden paths however we are confident on the golden path. Note: This may still be a valid end state for a feature... things like Beta features may well suffice at this level.
4 - We have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error/outlying scenarios. While more testing of the error/outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide.
5 - We have delivered all automated testing we believe is needed for this feature. The testing covers all golden path cases as well as all the error/outlying scenarios that make sense. We are not aware of any gaps in the testing at this time. No manual testing is required to verify this feature.
Based on your answer above, for any answer other than a 4 or 5 please provide details of what drove your answer. Please be aware, it may be perfectly reasonable in some scenarios to deliver with any value above. We may accept no automated testing is needed for some features, we may be happy with low levels of testing on samples for instance so please don't feel the need to drive to a 5. We need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid. What are the gaps, what is the risk etc. Please also provide links to the follow on work that is needed to close the gaps (should you deem it needed)
|
non_process
|
feature test summary test strategy test strategy followed what was done for the addition of a very similar property that allowed the sharing of connections on direct lookups test strategy what functionality is new or modified by this feature feature adds a boolean attribute enablecontainerauthfordirectlookups to existing connectionmanager config element if the value of the new attrib is true when a datasource is looked up w o injection or resource ref the default authentication mechanism will be container instead of application prior to this feature the only way to change the autheniticaton was via a resource ref what are the positive and negative tests for that functionality since the default behavior is unchanged ie application auth and we have tests that perform direct lookups as part of other tests the negative test path is already covered for the positive a test was added to the com ibm ws jdbc fat test suite that sets the new attrib to true performs a lookup obtains a connection and examines the credentials used by the database to make certain the specified container auth user was used to login and not the default what manual tests are there if any none confidence level collectively as a team you need to assess your confidence in the testing delivered based on the values below this should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole please indicate your confidence in the testing up to and including fat delivered with this feature by selecting one of these values no automated testing delivered we have minimal automated coverage of the feature including golden paths there is a relatively high risk that defects or issues could be found in this feature we have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here error outlying scenarios are not really covered there are likely risks that issues may exist in the golden paths we have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error outlying scenarios there is a risk when the feature is used outside the golden paths however we are confident on the golden path note this may still be a valid end state for a feature things like beta features may well suffice at this level we have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error outlying scenarios while more testing of the error outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide we have delivered all automated testing we believe is needed for this feature the testing covers all golden path cases as well as all the error outlying scenarios that make sense we are not aware of any gaps in the testing at this time no manual testing is required to verify this feature based on your answer above for any answer other than a or please provide details of what drove your answer please be aware it may be perfectly reasonable in some scenarios to deliver with any value above we may accept no automated testing is needed for some features we may be happy with low levels of testing on samples for instance so please don t feel the need to drive to a we need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid what are the gaps what is the risk etc please also provide links to the follow on work that is needed to close the gaps should you deem it needed
| 0
|
15,747
| 19,911,552,479
|
IssuesEvent
|
2022-01-25 17:39:54
|
input-output-hk/high-assurance-legacy
|
https://api.github.com/repos/input-output-hk/high-assurance-legacy
|
closed
|
Add Isabelle export of Haskell-embedded ♮-calculus code
|
type: enhancement language: isabelle language: haskell topic: process calculus
|
Our goal is to implement an interpretation that turns a ♮-calculus program into Isabelle source code that represents this program using the Isabelle embedding of the calculus.
|
1.0
|
Add Isabelle export of Haskell-embedded ♮-calculus code - Our goal is to implement an interpretation that turns a ♮-calculus program into Isabelle source code that represents this program using the Isabelle embedding of the calculus.
|
process
|
add isabelle export of haskell embedded ♮ calculus code our goal is to implement an interpretation that turns a ♮ calculus program into isabelle source code that represents this program using the isabelle embedding of the calculus
| 1
|
18,801
| 24,702,423,445
|
IssuesEvent
|
2022-10-19 16:13:41
|
NREL/buildstockbatch
|
https://api.github.com/repos/NREL/buildstockbatch
|
closed
|
Timeseries processing doesn't get triggered for ComStock projects
|
bug postprocessing eagle
|
**Describe the bug**
The timeseries postprocessing isn't triggered in ComStock due to misalignment of the workflow generators for commercial and residential (see #222). Basically https://github.com/NREL/buildstockbatch/blob/1cca3c82fba1d013f7f81bbad4ba05a4f6521d11/buildstockbatch/base.py#L545 is not triggered by the commercial workflow yaml which looks like this:
```yml
workflow_generator:
type: commercial_default
args:
reporting_measures:
- measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results
arguments:
building_summary_section: true
annual_overview_section: true
monthly_overview_section: true
utility_bills_rates_section: true
envelope_section_section: true
space_type_breakdown_section: true
space_type_details_section: true
interior_lighting_section: true
plug_loads_section: true
exterior_light_section: true
water_use_section: true
hvac_load_profile: true
zone_condition_section: true
zone_summary_section: true
zone_equipment_detail_section: true
air_loops_detail_section: true
plant_loops_detail_section: true
outdoor_air_section: true
cost_summary_section: true
source_energy_section: true
schedules_overview_section: true
- measure_dir_name: SimulationOutputReport
- measure_dir_name: TimeseriesCSVExport
arguments:
reporting_frequency: "Timestep"
inc_output_variables: false
- measure_dir_name: comstock_sensitivity_reports
- measure_dir_name: qoi_report
- measure_dir_name: la_100_qaqc
arguments:
run_qaqc: false
- measure_dir_name: simulation_settings_check
arguments:
run_sim_settings_checks: true
```
**To Reproduce**
Steps to reproduce the behavior:
1. Example yaml from @mdahlhausen:
```yml
schema_version: '0.3'
buildstock_directory: /lustre/eaglefs/projects/enduse/comstock/repos/comstock
project_directory: national
output_directory: /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/
weather_files_path: /lustre/eaglefs/projects/enduse/comstock/weather/Fort_Collins_County_IDs_2016.zip
sampler:
type: precomputed
args:
sample_file: /lustre/eaglefs/projects/enduse/comstock/samples/buildstock_wwr_2016_test.csv
eagle:
account: enduse
n_jobs: 1
minutes_per_sim: 120
postprocessing:
time: 45
postprocessing:
keep_individual_timeseries: true
aws:
region_name: 'us-west-2'
s3:
bucket: eulp
prefix: simulation_output/regional_runs/comstock/com_test_17_2016/
athena:
glue_service_role: service-role/AWSGlueServiceRole-default
database_name: enduse
max_crawling_time: 600 # Time to wait for the crawler to complete before aborting it
baseline:
n_buildings_represented: 30
custom_gems: True
os_version: eulp_com_v12
os_sha: 3472e8b799
workflow_generator:
type: commercial_default
args:
reporting_measures:
- measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results
arguments:
building_summary_section: true
annual_overview_section: true
monthly_overview_section: true
utility_bills_rates_section: true
envelope_section_section: true
space_type_breakdown_section: true
space_type_details_section: true
interior_lighting_section: true
plug_loads_section: true
exterior_light_section: true
water_use_section: true
hvac_load_profile: true
zone_condition_section: true
zone_summary_section: true
zone_equipment_detail_section: true
air_loops_detail_section: true
plant_loops_detail_section: true
outdoor_air_section: true
cost_summary_section: true
source_energy_section: true
schedules_overview_section: true
- measure_dir_name: SimulationOutputReport
- measure_dir_name: TimeseriesCSVExport
arguments:
reporting_frequency: "Timestep"
inc_output_variables: false
- measure_dir_name: comstock_sensitivity_reports
- measure_dir_name: qoi_report
- measure_dir_name: la_100_qaqc
arguments:
run_qaqc: false
- measure_dir_name: simulation_settings_check
arguments:
run_sim_settings_checks: true
```
2. That's about it.
**Expected behavior**
I want (but don't expect given the code) for the timeseries postprocessing to run. But it doesn't.
**Logs**
postprocessing.out
```
begin eagle_postprocessing.sh
Job ID: 6997691
Hostname: r2i7n29
UPLOADONLY: False
MEMORY: 85248
head node
r2i7n29
workers
r8i0n[11,15]
r8i0n11: Warning: Permanently added 'r8i0n11,10.148.8.126' (ECDSA) to the list of known hosts.
r8i0n15: Warning: Permanently added 'r8i0n15,10.148.8.134' (ECDSA) to the list of known hosts.
r8i0n15: total used free shared buff/cache available
r8i0n15: Mem: 92G 2.9G 84G 5.1G 5.2G 84G
r8i0n15: Swap: 0B 0B 0B
r8i0n11: total used free shared buff/cache available
r8i0n11: Mem: 92G 2.9G 85G 3.9G 4.1G 85G
r8i0n11: Swap: 0B 0B 0B
DEBUG:2021-06-15 13:01:24:buildstockbatch.base:Using OpenStudio version: eulp_com_v12 with SHA: 3472e8b799
DEBUG:2021-06-15 13:01:24:__main__:Output directory = /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016
DEBUG:2021-06-15 13:01:24:__main__:Starting postprocessing
INFO:2021-06-15 13:01:28:buildstockbatch.postprocessing:Creating results_df.
INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Writing /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/results/results_csvs/results_up00.csv.gz
INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Uploading the parquet files to s3
INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Upload to S3 completed. The files are uploaded to: eulp/simulation_output/regional_runs/comstock/com_test_17_2016/com_test_17_2016/
INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Creating Athena tables using glue crawler
INFO:2021-06-15 13:01:39:buildstockbatch.postprocessing:Deleting existing tables in db enduse: ['_com_test_17_2016']. And creating new ones.
INFO:2021-06-15 13:01:51:buildstockbatch.postprocessing:Crawler started
INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:Crawler has completed running. It is STOPPING.
INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:TablesCreated: 1 TablesUpdated: 0 TablesDeleted: 0
INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Crawler stopped. Deleting it now.
INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Removing results_job*.json.gz
real 5m45.335s
user 0m1.451s
sys 0m2.137s
```
**Platform (please complete the following information):**
- Simulation platform: eagle
- BuildStockBatch version, branch, or sha: 1cca3c82fba1d013f7f81bbad4ba05a4f6521d11
- resstock or comstock repo version, branch, or sha: no clue
- Local Desktop OS: n/a
**Additional context**
This is really a issue with the workflow generator schema / specification but I have no idea what the proper solution is there. Instead i'm going to role a crappy patch for this that will get us through. In general I'm not a huge fan of how timeseries postprocessing is (or really isn't) specified in the yaml but that's also a different issue for a different day.
|
1.0
|
Timeseries processing doesn't get triggered for ComStock projects - **Describe the bug**
The timeseries postprocessing isn't triggered in ComStock due to misalignment of the workflow generators for commercial and residential (see #222). Basically https://github.com/NREL/buildstockbatch/blob/1cca3c82fba1d013f7f81bbad4ba05a4f6521d11/buildstockbatch/base.py#L545 is not triggered by the commercial workflow yaml which looks like this:
```yml
workflow_generator:
type: commercial_default
args:
reporting_measures:
- measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results
arguments:
building_summary_section: true
annual_overview_section: true
monthly_overview_section: true
utility_bills_rates_section: true
envelope_section_section: true
space_type_breakdown_section: true
space_type_details_section: true
interior_lighting_section: true
plug_loads_section: true
exterior_light_section: true
water_use_section: true
hvac_load_profile: true
zone_condition_section: true
zone_summary_section: true
zone_equipment_detail_section: true
air_loops_detail_section: true
plant_loops_detail_section: true
outdoor_air_section: true
cost_summary_section: true
source_energy_section: true
schedules_overview_section: true
- measure_dir_name: SimulationOutputReport
- measure_dir_name: TimeseriesCSVExport
arguments:
reporting_frequency: "Timestep"
inc_output_variables: false
- measure_dir_name: comstock_sensitivity_reports
- measure_dir_name: qoi_report
- measure_dir_name: la_100_qaqc
arguments:
run_qaqc: false
- measure_dir_name: simulation_settings_check
arguments:
run_sim_settings_checks: true
```
**To Reproduce**
Steps to reproduce the behavior:
1. Example yaml from @mdahlhausen:
```yml
schema_version: '0.3'
buildstock_directory: /lustre/eaglefs/projects/enduse/comstock/repos/comstock
project_directory: national
output_directory: /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/
weather_files_path: /lustre/eaglefs/projects/enduse/comstock/weather/Fort_Collins_County_IDs_2016.zip
sampler:
type: precomputed
args:
sample_file: /lustre/eaglefs/projects/enduse/comstock/samples/buildstock_wwr_2016_test.csv
eagle:
account: enduse
n_jobs: 1
minutes_per_sim: 120
postprocessing:
time: 45
postprocessing:
keep_individual_timeseries: true
aws:
region_name: 'us-west-2'
s3:
bucket: eulp
prefix: simulation_output/regional_runs/comstock/com_test_17_2016/
athena:
glue_service_role: service-role/AWSGlueServiceRole-default
database_name: enduse
max_crawling_time: 600 # Time to wait for the crawler to complete before aborting it
baseline:
n_buildings_represented: 30
custom_gems: True
os_version: eulp_com_v12
os_sha: 3472e8b799
workflow_generator:
type: commercial_default
args:
reporting_measures:
- measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results
arguments:
building_summary_section: true
annual_overview_section: true
monthly_overview_section: true
utility_bills_rates_section: true
envelope_section_section: true
space_type_breakdown_section: true
space_type_details_section: true
interior_lighting_section: true
plug_loads_section: true
exterior_light_section: true
water_use_section: true
hvac_load_profile: true
zone_condition_section: true
zone_summary_section: true
zone_equipment_detail_section: true
air_loops_detail_section: true
plant_loops_detail_section: true
outdoor_air_section: true
cost_summary_section: true
source_energy_section: true
schedules_overview_section: true
- measure_dir_name: SimulationOutputReport
- measure_dir_name: TimeseriesCSVExport
arguments:
reporting_frequency: "Timestep"
inc_output_variables: false
- measure_dir_name: comstock_sensitivity_reports
- measure_dir_name: qoi_report
- measure_dir_name: la_100_qaqc
arguments:
run_qaqc: false
- measure_dir_name: simulation_settings_check
arguments:
run_sim_settings_checks: true
```
2. That's about it.
**Expected behavior**
I want (but don't expect given the code) for the timeseries postprocessing to run. But it doesn't.
**Logs**
postprocessing.out
```
begin eagle_postprocessing.sh
Job ID: 6997691
Hostname: r2i7n29
UPLOADONLY: False
MEMORY: 85248
head node
r2i7n29
workers
r8i0n[11,15]
r8i0n11: Warning: Permanently added 'r8i0n11,10.148.8.126' (ECDSA) to the list of known hosts.
r8i0n15: Warning: Permanently added 'r8i0n15,10.148.8.134' (ECDSA) to the list of known hosts.
r8i0n15: total used free shared buff/cache available
r8i0n15: Mem: 92G 2.9G 84G 5.1G 5.2G 84G
r8i0n15: Swap: 0B 0B 0B
r8i0n11: total used free shared buff/cache available
r8i0n11: Mem: 92G 2.9G 85G 3.9G 4.1G 85G
r8i0n11: Swap: 0B 0B 0B
DEBUG:2021-06-15 13:01:24:buildstockbatch.base:Using OpenStudio version: eulp_com_v12 with SHA: 3472e8b799
DEBUG:2021-06-15 13:01:24:__main__:Output directory = /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016
DEBUG:2021-06-15 13:01:24:__main__:Starting postprocessing
INFO:2021-06-15 13:01:28:buildstockbatch.postprocessing:Creating results_df.
INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Writing /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/results/results_csvs/results_up00.csv.gz
INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Uploading the parquet files to s3
INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Upload to S3 completed. The files are uploaded to: eulp/simulation_output/regional_runs/comstock/com_test_17_2016/com_test_17_2016/
INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Creating Athena tables using glue crawler
INFO:2021-06-15 13:01:39:buildstockbatch.postprocessing:Deleting existing tables in db enduse: ['_com_test_17_2016']. And creating new ones.
INFO:2021-06-15 13:01:51:buildstockbatch.postprocessing:Crawler started
INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:Crawler has completed running. It is STOPPING.
INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:TablesCreated: 1 TablesUpdated: 0 TablesDeleted: 0
INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Crawler stopped. Deleting it now.
INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Removing results_job*.json.gz
real 5m45.335s
user 0m1.451s
sys 0m2.137s
```
**Platform (please complete the following information):**
- Simulation platform: eagle
- BuildStockBatch version, branch, or sha: 1cca3c82fba1d013f7f81bbad4ba05a4f6521d11
- resstock or comstock repo version, branch, or sha: no clue
- Local Desktop OS: n/a
**Additional context**
This is really a issue with the workflow generator schema / specification but I have no idea what the proper solution is there. Instead i'm going to role a crappy patch for this that will get us through. In general I'm not a huge fan of how timeseries postprocessing is (or really isn't) specified in the yaml but that's also a different issue for a different day.
|
process
|
timeseries processing doesn t get triggered for comstock projects describe the bug the timeseries postprocessing isn t triggered in comstock due to misalignment of the workflow generators for commercial and residential see basically is not triggered by the commercial workflow yaml which looks like this yml workflow generator type commercial default args reporting measures measure dir name openstudio results arguments building summary section true annual overview section true monthly overview section true utility bills rates section true envelope section section true space type breakdown section true space type details section true interior lighting section true plug loads section true exterior light section true water use section true hvac load profile true zone condition section true zone summary section true zone equipment detail section true air loops detail section true plant loops detail section true outdoor air section true cost summary section true source energy section true schedules overview section true measure dir name simulationoutputreport measure dir name timeseriescsvexport arguments reporting frequency timestep inc output variables false measure dir name comstock sensitivity reports measure dir name qoi report measure dir name la qaqc arguments run qaqc false measure dir name simulation settings check arguments run sim settings checks true to reproduce steps to reproduce the behavior example yaml from mdahlhausen yml schema version buildstock directory lustre eaglefs projects enduse comstock repos comstock project directory national output directory lustre eaglefs projects enduse comstock com test weather files path lustre eaglefs projects enduse comstock weather fort collins county ids zip sampler type precomputed args sample file lustre eaglefs projects enduse comstock samples buildstock wwr test csv eagle account enduse n jobs minutes per sim postprocessing time postprocessing keep individual timeseries true aws region name us west bucket eulp prefix simulation output regional runs comstock com test athena glue service role service role awsglueservicerole default database name enduse max crawling time time to wait for the crawler to complete before aborting it baseline n buildings represented custom gems true os version eulp com os sha workflow generator type commercial default args reporting measures measure dir name openstudio results arguments building summary section true annual overview section true monthly overview section true utility bills rates section true envelope section section true space type breakdown section true space type details section true interior lighting section true plug loads section true exterior light section true water use section true hvac load profile true zone condition section true zone summary section true zone equipment detail section true air loops detail section true plant loops detail section true outdoor air section true cost summary section true source energy section true schedules overview section true measure dir name simulationoutputreport measure dir name timeseriescsvexport arguments reporting frequency timestep inc output variables false measure dir name comstock sensitivity reports measure dir name qoi report measure dir name la qaqc arguments run qaqc false measure dir name simulation settings check arguments run sim settings checks true that s about it expected behavior i want but don t expect given the code for the timeseries postprocessing to run but it doesn t logs postprocessing out begin eagle postprocessing sh job id hostname uploadonly false memory head node workers warning permanently added ecdsa to the list of known hosts warning permanently added ecdsa to the list of known hosts total used free shared buff cache available mem swap total used free shared buff cache available mem swap debug buildstockbatch base using openstudio version eulp com with sha debug main output directory lustre eaglefs projects enduse comstock com test debug main starting postprocessing info buildstockbatch postprocessing creating results df info buildstockbatch postprocessing writing lustre eaglefs projects enduse comstock com test results results csvs results csv gz info buildstockbatch postprocessing uploading the parquet files to info buildstockbatch postprocessing upload to completed the files are uploaded to eulp simulation output regional runs comstock com test com test info buildstockbatch postprocessing creating athena tables using glue crawler info buildstockbatch postprocessing deleting existing tables in db enduse and creating new ones info buildstockbatch postprocessing crawler started info buildstockbatch postprocessing crawler has completed running it is stopping info buildstockbatch postprocessing tablescreated tablesupdated tablesdeleted info buildstockbatch postprocessing crawler stopped deleting it now info buildstockbatch postprocessing removing results job json gz real user sys platform please complete the following information simulation platform eagle buildstockbatch version branch or sha resstock or comstock repo version branch or sha no clue local desktop os n a additional context this is really a issue with the workflow generator schema specification but i have no idea what the proper solution is there instead i m going to role a crappy patch for this that will get us through in general i m not a huge fan of how timeseries postprocessing is or really isn t specified in the yaml but that s also a different issue for a different day
| 1
|
92,717
| 8,376,569,481
|
IssuesEvent
|
2018-10-05 20:18:27
|
dojot/dojot
|
https://api.github.com/repos/dojot/dojot
|
closed
|
[GUI] Map view - centralization
|
Priority:High Status:ToTest Team:Frontend Type:Bug
|
Map centralized in the region of Pradópolis but devices are in Campinas.


**Affected Version:** 0.3.0-nightly_20180912
|
1.0
|
[GUI] Map view - centralization - Map centralized in the region of Pradópolis but devices are in Campinas.


**Affected Version:** 0.3.0-nightly_20180912
|
non_process
|
map view centralization map centralized in the region of pradópolis but devices are in campinas affected version nightly
| 0
|
422,205
| 28,377,116,692
|
IssuesEvent
|
2023-04-12 21:55:22
|
EECE3093C/team-project-weather
|
https://api.github.com/repos/EECE3093C/team-project-weather
|
opened
|
Label all iteration 4 issues as functional/non-functional
|
documentation
|
Ensure all issues are labeled as such.
This is a functional requirement.
|
1.0
|
Label all iteration 4 issues as functional/non-functional - Ensure all issues are labeled as such.
This is a functional requirement.
|
non_process
|
label all iteration issues as functional non functional ensure all issues are labeled as such this is a functional requirement
| 0
|
136,978
| 11,094,017,108
|
IssuesEvent
|
2019-12-16 04:30:53
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
CIS Scan - Failed nodes not reported in the API when the status of test is "mixed"
|
[zube]: To Test area/scan-tool kind/bug-qa team/ca
|
**What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- On a custom cluster - all roles - 1 node, run CIS Scans.
- For a test which has the state as "mixed", API report does not list the nodes it has failed on
**Expected Result:**
The report in API should report the nodes which have failed when the state of the test is "mixed"
**Other details that may be helpful:**
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master - head
- Installation option (single install/HA): single
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): custom
- Kubernetes version (use `kubectl version`):
```
1.16.3
```
|
1.0
|
CIS Scan - Failed nodes not reported in the API when the status of test is "mixed" - **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- On a custom cluster - all roles - 1 node, run CIS Scans.
- For a test which has the state as "mixed", API report does not list the nodes it has failed on
**Expected Result:**
The report in API should report the nodes which have failed when the state of the test is "mixed"
**Other details that may be helpful:**
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master - head
- Installation option (single install/HA): single
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): custom
- Kubernetes version (use `kubectl version`):
```
1.16.3
```
|
non_process
|
cis scan failed nodes not reported in the api when the status of test is mixed what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible on a custom cluster all roles node run cis scans for a test which has the state as mixed api report does not list the nodes it has failed on expected result the report in api should report the nodes which have failed when the state of the test is mixed other details that may be helpful environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui master head installation option single install ha single if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported custom kubernetes version use kubectl version
| 0
|
17,207
| 22,792,178,938
|
IssuesEvent
|
2022-07-10 06:52:42
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
opened
|
[processor/metricstransform] Remove empty metrics
|
priority:p3 processor/metricstransform
|
Some operations, e.g. `remove_label_values`, can leave metric without data points. That kind of metrics can be considered invalid and should be removed
|
1.0
|
[processor/metricstransform] Remove empty metrics - Some operations, e.g. `remove_label_values`, can leave metric without data points. That kind of metrics can be considered invalid and should be removed
|
process
|
remove empty metrics some operations e g remove label values can leave metric without data points that kind of metrics can be considered invalid and should be removed
| 1
|
47,018
| 2,971,267,420
|
IssuesEvent
|
2015-07-14 04:57:56
|
nabu-catalog/nabu
|
https://api.github.com/repos/nabu-catalog/nabu
|
closed
|
Delete items and collections
|
2015 Priorities Priority
|
We need the ability to delete collections even if they have content, and also to delete items if they have content. At the moment we need to delete each essence file before we can delete an item and each item before we can delete a collection.
|
1.0
|
Delete items and collections - We need the ability to delete collections even if they have content, and also to delete items if they have content. At the moment we need to delete each essence file before we can delete an item and each item before we can delete a collection.
|
non_process
|
delete items and collections we need the ability to delete collections even if they have content and also to delete items if they have content at the moment we need to delete each essence file before we can delete an item and each item before we can delete a collection
| 0
|
1,430
| 3,995,445,257
|
IssuesEvent
|
2016-05-10 15:32:35
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
opened
|
Wrong pass `this` for function constructor
|
AREA: client SYSTEM: resource processing TYPE: bug
|
Found in https://testcafe-hhhm.devexpress.com/history?timestamp=1462872434105&siteUrl=http://shareasale.com/
Script for reproduce:
```javascript
var Ext = {
toString: function () {
return 'Ext';
}
};
Ext.data = {
toString: function () {
return 'Ext.data';
}
};
Ext.data.Connection = function () {
console.log(this);
//Without proxing `this` equal Ext.data.Connection
//With proxing `this` equal the parent scope - Ext.data
};
Ext.data.Connection.toString = function () {
return 'Ext.data.Connection';
}
var obj = new Ext.data.Connection();
```
Possible cause - https://github.com/DevExpress/testcafe-hammerhead/pull/562
|
1.0
|
Wrong pass `this` for function constructor - Found in https://testcafe-hhhm.devexpress.com/history?timestamp=1462872434105&siteUrl=http://shareasale.com/
Script for reproduce:
```javascript
var Ext = {
toString: function () {
return 'Ext';
}
};
Ext.data = {
toString: function () {
return 'Ext.data';
}
};
Ext.data.Connection = function () {
console.log(this);
//Without proxing `this` equal Ext.data.Connection
//With proxing `this` equal the parent scope - Ext.data
};
Ext.data.Connection.toString = function () {
return 'Ext.data.Connection';
}
var obj = new Ext.data.Connection();
```
Possible cause - https://github.com/DevExpress/testcafe-hammerhead/pull/562
|
process
|
wrong pass this for function constructor found in script for reproduce javascript var ext tostring function return ext ext data tostring function return ext data ext data connection function console log this without proxing this equal ext data connection with proxing this equal the parent scope ext data ext data connection tostring function return ext data connection var obj new ext data connection possible cause
| 1
|
22,586
| 31,811,331,265
|
IssuesEvent
|
2023-09-13 17:03:10
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
reopened
|
Obsoletion request: RNA polyadenylation
|
RNA processes obsoletion ready MF_in_BP
|
Please provide as much information as you can:
Once https://github.com/geneontology/go-ontology/issues/24921
is done we can begin to obsolete
- [x] GO:0043631 RNA polyadenylation
- [x] GO:0006378 mRNA polyadenylation
- [x] GO:1900364 negative regulation of mRNA polyadenylation
- [x] GO:1900365 positive regulation of mRNA polyadenylation
- [x] GO:1900363 regulation of mRNA polyadenylation
- [x] GO:0043629 ncRNA polyadenylation
(I will make sure there are direct replacements for all of these)
- [x] GO:0043630 ncRNA polyadenylation involved in polyadenylation-dependent ncRNA catabolic process
replace by -> GO:0043634 polyadenylation-dependent ncRNA catabolic process
- [x] GO:0071050 sno(s)RNA polyadenylation ->
replace by GO:0071051 polyadenylation-dependent snoRNA 3'-end processing
Rehouse:
- [ ] GO:0097222 mitochondrial mRNA polyadenylation
- [ ] GO:0140235 RNA polyadenylation at postsynapse (this could probably move under cytoplasmic polyadenylation, need to check)
- [ ] GO:0039698 polyadenylation of viral mRNA by polymerase stuttering (this is something completely different nd ned rehoming)
* **GO term ID and Label**
* **Reason for deprecation** Put an x in the appropriate box:
- [x] The reason for obsoletion is that this term represents a molecular function.
* **"Replace by" term (ID and label)**
If all annotations can safely be moved to that term
* **"Consider" term(s) (ID and label)**
Suggestions for reannotation
* **Are there annotations to this term?**
- How many EXP:
* **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)**
* **Is this term in a subset? (check the AmiGO page for that term)**
* **Any other information**
----
Checklist for ontology editor
***Check term usage and metadata in Protégé***
- [ ] check term usage in the ontology
- [ ] check internal mappings: RHEA, EC, MetaCyc
- [ ] check subset usage
- [ ] check taxon constraints
***Check annotations***
- [x] create Annotation Review spreadsheet
https://docs.google.com/spreadsheets/d/1Dyb4jkJAqnHonfWw2-9u3zQcpAcVd5cd9p_I7ta2Muc/edit#gid=0
- [x] Annotation review ticket https://github.com/geneontology/go-annotation/issues/4553
- [x] find external mappings (via IEAS), include in obsoletion notice
- [x] tag curators in Annotation Review ticket. Responsible annotator for each group: https://github.com/geneontology/go-site/blob/master/metadata/group-contacts.csv
***Notification***
- [ ] create [obsoletion announcement](https://github.com/geneontology/go-announcements/issues/new?assignees=&labels=obsoletion&template=obsoletion-notice.md&title=Obsoletion+notice%3A+%5BGO+ID%3A+term+label%5D)
- [ ] announce to GO friends (go-friends@mailman.stanford.edu)
- [ ] paste the text in the ontology ticket
|
1.0
|
Obsoletion request: RNA polyadenylation - Please provide as much information as you can:
Once https://github.com/geneontology/go-ontology/issues/24921
is done we can begin to obsolete
- [x] GO:0043631 RNA polyadenylation
- [x] GO:0006378 mRNA polyadenylation
- [x] GO:1900364 negative regulation of mRNA polyadenylation
- [x] GO:1900365 positive regulation of mRNA polyadenylation
- [x] GO:1900363 regulation of mRNA polyadenylation
- [x] GO:0043629 ncRNA polyadenylation
(I will make sure there are direct replacements for all of these)
- [x] GO:0043630 ncRNA polyadenylation involved in polyadenylation-dependent ncRNA catabolic process
replace by -> GO:0043634 polyadenylation-dependent ncRNA catabolic process
- [x] GO:0071050 sno(s)RNA polyadenylation ->
replace by GO:0071051 polyadenylation-dependent snoRNA 3'-end processing
Rehouse:
- [ ] GO:0097222 mitochondrial mRNA polyadenylation
- [ ] GO:0140235 RNA polyadenylation at postsynapse (this could probably move under cytoplasmic polyadenylation, need to check)
- [ ] GO:0039698 polyadenylation of viral mRNA by polymerase stuttering (this is something completely different nd ned rehoming)
* **GO term ID and Label**
* **Reason for deprecation** Put an x in the appropriate box:
- [x] The reason for obsoletion is that this term represents a molecular function.
* **"Replace by" term (ID and label)**
If all annotations can safely be moved to that term
* **"Consider" term(s) (ID and label)**
Suggestions for reannotation
* **Are there annotations to this term?**
- How many EXP:
* **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)**
* **Is this term in a subset? (check the AmiGO page for that term)**
* **Any other information**
----
Checklist for ontology editor
***Check term usage and metadata in Protégé***
- [ ] check term usage in the ontology
- [ ] check internal mappings: RHEA, EC, MetaCyc
- [ ] check subset usage
- [ ] check taxon constraints
***Check annotations***
- [x] create Annotation Review spreadsheet
https://docs.google.com/spreadsheets/d/1Dyb4jkJAqnHonfWw2-9u3zQcpAcVd5cd9p_I7ta2Muc/edit#gid=0
- [x] Annotation review ticket https://github.com/geneontology/go-annotation/issues/4553
- [x] find external mappings (via IEAS), include in obsoletion notice
- [x] tag curators in Annotation Review ticket. Responsible annotator for each group: https://github.com/geneontology/go-site/blob/master/metadata/group-contacts.csv
***Notification***
- [ ] create [obsoletion announcement](https://github.com/geneontology/go-announcements/issues/new?assignees=&labels=obsoletion&template=obsoletion-notice.md&title=Obsoletion+notice%3A+%5BGO+ID%3A+term+label%5D)
- [ ] announce to GO friends (go-friends@mailman.stanford.edu)
- [ ] paste the text in the ontology ticket
|
process
|
obsoletion request rna polyadenylation please provide as much information as you can once is done we can begin to obsolete go rna polyadenylation go mrna polyadenylation go negative regulation of mrna polyadenylation go positive regulation of mrna polyadenylation go regulation of mrna polyadenylation go ncrna polyadenylation i will make sure there are direct replacements for all of these go ncrna polyadenylation involved in polyadenylation dependent ncrna catabolic process replace by go polyadenylation dependent ncrna catabolic process go sno s rna polyadenylation replace by go polyadenylation dependent snorna end processing rehouse go mitochondrial mrna polyadenylation go rna polyadenylation at postsynapse this could probably move under cytoplasmic polyadenylation need to check go polyadenylation of viral mrna by polymerase stuttering this is something completely different nd ned rehoming go term id and label reason for deprecation put an x in the appropriate box the reason for obsoletion is that this term represents a molecular function replace by term id and label if all annotations can safely be moved to that term consider term s id and label suggestions for reannotation are there annotations to this term how many exp are there mappings and cross references to this term interpro keywords check quickgo cross references section is this term in a subset check the amigo page for that term any other information checklist for ontology editor check term usage and metadata in protégé check term usage in the ontology check internal mappings rhea ec metacyc check subset usage check taxon constraints check annotations create annotation review spreadsheet annotation review ticket find external mappings via ieas include in obsoletion notice tag curators in annotation review ticket responsible annotator for each group notification create announce to go friends go friends mailman stanford edu paste the text in the ontology ticket
| 1
|
18,943
| 24,904,581,998
|
IssuesEvent
|
2022-10-29 04:28:53
|
shirou/gopsutil
|
https://api.github.com/repos/shirou/gopsutil
|
closed
|
Windows Process.Name() getFromSnapProcess() performance
|
os:windows package:process performance
|
Currently `Process.Name()` on Windows calls the `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` syscall to get it's ppid and name:
https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L321-L322
https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L830-L831
It seems `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` will return all processes in the system despite passing the ProcessID parameter.
Although returning a snapshot of all processes is quite fast most of the time (<10ms), it can sometimes be slow (>100ms, sometimes even >1000ms).
Can we change the flags like below according to the documentation:
(not so sure about the exact flags, especially `TH32CS_SNAPMODULE32`)
```
func getFromSnapProcess(pid int32) (int32, int32, string, error) {
// https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot
// To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero.
// Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again,
// specifying its process identifier and the TH32CS_SNAPHEAPLIST or TH32_SNAPMODULE value.
snap, err := windows.CreateToolhelp32Snapshot(windows.TH32CS_SNAPHEAPLIST|windows.TH32CS_SNAPMODULE|windows.TH32CS_SNAPMODULE32, uint32(pid))
if err != nil {
return 0, 0, "", err
}
defer windows.CloseHandle(snap)
var pe32 windows.ProcessEntry32
pe32.Size = uint32(unsafe.Sizeof(pe32))
if err = windows.Process32First(snap, &pe32); err != nil && err != syscall.ERROR_NO_MORE_FILES {
return 0, 0, "", err
}
for {
if pe32.ProcessID == uint32(pid) {
szexe := windows.UTF16ToString(pe32.ExeFile[:])
return int32(pe32.ParentProcessID), int32(pe32.Threads), szexe, nil
}
if err = windows.Process32Next(snap, &pe32); err != nil {
break
}
}
return 0, 0, "", fmt.Errorf("couldn't find pid: %d", pid)
}
```
or switch to `NtQuerySystemInformation`?
## refs
https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot:
- TH32CS_SNAPPROCESS 0x00000002 | Includes all processes in the system in the snapshot.
> To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero.
> Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again, specifying its process identifier and the **TH32CS_SNAPHEAPLIST** or **TH32_SNAPMODULE** value.
## see also
- https://github.com/Maximus5/ConEmu/issues/885
- https://sudonull.com/post/61039-Speeding-up-the-transfer-of-processes-and-threads-in-Windows-Infopulse-Ukraine-Blog
- https://learn.microsoft.com/en-us/windows/win32/api/winternl/nf-winternl-ntquerysysteminformation
|
1.0
|
Windows Process.Name() getFromSnapProcess() performance - Currently `Process.Name()` on Windows calls the `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` syscall to get it's ppid and name:
https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L321-L322
https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L830-L831
It seems `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` will return all processes in the system despite passing the ProcessID parameter.
Although returning a snapshot of all processes is quite fast most of the time (<10ms), it can sometimes be slow (>100ms, sometimes even >1000ms).
Can we change the flags like below according to the documentation:
(not so sure about the exact flags, especially `TH32CS_SNAPMODULE32`)
```
func getFromSnapProcess(pid int32) (int32, int32, string, error) {
// https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot
// To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero.
// Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again,
// specifying its process identifier and the TH32CS_SNAPHEAPLIST or TH32_SNAPMODULE value.
snap, err := windows.CreateToolhelp32Snapshot(windows.TH32CS_SNAPHEAPLIST|windows.TH32CS_SNAPMODULE|windows.TH32CS_SNAPMODULE32, uint32(pid))
if err != nil {
return 0, 0, "", err
}
defer windows.CloseHandle(snap)
var pe32 windows.ProcessEntry32
pe32.Size = uint32(unsafe.Sizeof(pe32))
if err = windows.Process32First(snap, &pe32); err != nil && err != syscall.ERROR_NO_MORE_FILES {
return 0, 0, "", err
}
for {
if pe32.ProcessID == uint32(pid) {
szexe := windows.UTF16ToString(pe32.ExeFile[:])
return int32(pe32.ParentProcessID), int32(pe32.Threads), szexe, nil
}
if err = windows.Process32Next(snap, &pe32); err != nil {
break
}
}
return 0, 0, "", fmt.Errorf("couldn't find pid: %d", pid)
}
```
or switch to `NtQuerySystemInformation`?
## refs
https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot:
- TH32CS_SNAPPROCESS 0x00000002 | Includes all processes in the system in the snapshot.
> To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero.
> Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again, specifying its process identifier and the **TH32CS_SNAPHEAPLIST** or **TH32_SNAPMODULE** value.
## see also
- https://github.com/Maximus5/ConEmu/issues/885
- https://sudonull.com/post/61039-Speeding-up-the-transfer-of-processes-and-threads-in-Windows-Infopulse-Ukraine-Blog
- https://learn.microsoft.com/en-us/windows/win32/api/winternl/nf-winternl-ntquerysysteminformation
|
process
|
windows process name getfromsnapprocess performance currently process name on windows calls the snapprocess pid syscall to get it s ppid and name it seems snapprocess pid will return all processes in the system despite passing the processid parameter although returning a snapshot of all processes is quite fast most of the time sometimes even can we change the flags like below according to the documentation not so sure about the exact flags especially func getfromsnapprocess pid string error to enumerate the heap or module states for all processes specify snapall and set to zero then for each additional process in the snapshot call again specifying its process identifier and the snapheaplist or snapmodule value snap err windows windows snapheaplist windows snapmodule windows pid if err nil return err defer windows closehandle snap var windows size unsafe sizeof if err windows snap err nil err syscall error no more files return err for if processid pid szexe windows exefile return parentprocessid threads szexe nil if err windows snap err nil break return fmt errorf couldn t find pid d pid or switch to ntquerysysteminformation refs snapprocess includes all processes in the system in the snapshot to enumerate the heap or module states for all processes specify snapall and set to zero then for each additional process in the snapshot call again specifying its process identifier and the snapheaplist or snapmodule value see also
| 1
|
7,209
| 10,343,374,395
|
IssuesEvent
|
2019-09-04 08:49:21
|
Hurence/logisland
|
https://api.github.com/repos/Hurence/logisland
|
closed
|
Add new processor URLDecode
|
processor
|
# Expected behavior and actual behavior.
When a field contains a URL, it can be needed to decode special chars encoded in URL (in UT-8 charset for instance) such that it does not appear encoded and mess up further user data processing.
The need is then to be able to specify a list of one or more fields subject to URL decoding and a charset (default being UTF-8).
# Steps to reproduce the problem.
# Specifications like the version of the project, operating system, or hardware.
|
1.0
|
Add new processor URLDecode - # Expected behavior and actual behavior.
When a field contains a URL, it can be needed to decode special chars encoded in URL (in UT-8 charset for instance) such that it does not appear encoded and mess up further user data processing.
The need is then to be able to specify a list of one or more fields subject to URL decoding and a charset (default being UTF-8).
# Steps to reproduce the problem.
# Specifications like the version of the project, operating system, or hardware.
|
process
|
add new processor urldecode expected behavior and actual behavior when a field contains a url it can be needed to decode special chars encoded in url in ut charset for instance such that it does not appear encoded and mess up further user data processing the need is then to be able to specify a list of one or more fields subject to url decoding and a charset default being utf steps to reproduce the problem specifications like the version of the project operating system or hardware
| 1
|
21,538
| 29,837,211,788
|
IssuesEvent
|
2023-06-19 00:37:44
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Hibrido / Belo Horizonte, Minas Gerais, Brazil] Fullstack Developer (.NET/Javascript) na Coodesh
|
SALVADOR JAVASCRIPT FULL-STACK CSS3 MVC PLENO SQL GIT BOOTSTRAP WEB API ENTITY FRAMEWORK MOBILE REQUISITOS ASP.NET PROCESSOS GITHUB UMA C QUALIDADE MANUTENÇÃO ALOCADO Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Revelar RH está em busca de Fullstack Developer para compor seu time!</p>
<p>Responsabilidades:</p>
<ul>
<li>Desenvolvimento e acompanhamento de manutenção evolutivas e corretivas em sistemas desenvolvidos em .NET;</li>
<li>Propor soluções e desenvolver novas tecnologias para melhoria dos sistemas;</li>
<li>Manter a qualidade do código, versionamento, testes, deploy e participação na priorização de funcionalidades.</li>
</ul>
<p>Horário de segunda à sexta das 8h às 17h.</p>
## Revelar RH:
<p>A Revelar RH é uma empresa inovadora que busca alinhar os objetivos das Empresas que têm carência de profissionais qualificados, com os das pessoas que estão em busca de uma oportunidades no mercado de trabalho. Nosso trabalho visa o aprimoramento e racionalização de processos voltados à Gestão de Pessoas, sendo realizado de forma personalizada para cada cliente. Atualmente atendemos várias empresas de tecnologia e softwares house. </p></p>
## Habilidades:
- Asp.Net MVC
- C#
- Arquitetura de Projetos Mobile
## Local:
Belo Horizonte, Minas Gerais, Brazil
## Requisitos:
- Superior completo ou em curso de Análise e Desenvolvimento de Sistemas, Engenharia da computação ou áreas afins;
- Experiência e conhecimento correspondente ao perfil pleno/sênior;
- Experiência com HTML5, CSS3, Bootstrap;
- Experiência com SQL SERVER;
- Experiência com GIT;
- Experiência com ASP.NET MVC, WEB API, C# e Entity framework.
## Benefícios:
- Plano de saúde;
- Plano odontológico;
- Seguro de vida;
- Cartão de benefícios Swile (alimentação, auxílio internet, desconto em academia).
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (.NET/Javascript) na Revelar RH](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Categoria
Full-Stack
|
1.0
|
[Hibrido / Belo Horizonte, Minas Gerais, Brazil] Fullstack Developer (.NET/Javascript) na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Revelar RH está em busca de Fullstack Developer para compor seu time!</p>
<p>Responsabilidades:</p>
<ul>
<li>Desenvolvimento e acompanhamento de manutenção evolutivas e corretivas em sistemas desenvolvidos em .NET;</li>
<li>Propor soluções e desenvolver novas tecnologias para melhoria dos sistemas;</li>
<li>Manter a qualidade do código, versionamento, testes, deploy e participação na priorização de funcionalidades.</li>
</ul>
<p>Horário de segunda à sexta das 8h às 17h.</p>
## Revelar RH:
<p>A Revelar RH é uma empresa inovadora que busca alinhar os objetivos das Empresas que têm carência de profissionais qualificados, com os das pessoas que estão em busca de uma oportunidades no mercado de trabalho. Nosso trabalho visa o aprimoramento e racionalização de processos voltados à Gestão de Pessoas, sendo realizado de forma personalizada para cada cliente. Atualmente atendemos várias empresas de tecnologia e softwares house. </p></p>
## Habilidades:
- Asp.Net MVC
- C#
- Arquitetura de Projetos Mobile
## Local:
Belo Horizonte, Minas Gerais, Brazil
## Requisitos:
- Superior completo ou em curso de Análise e Desenvolvimento de Sistemas, Engenharia da computação ou áreas afins;
- Experiência e conhecimento correspondente ao perfil pleno/sênior;
- Experiência com HTML5, CSS3, Bootstrap;
- Experiência com SQL SERVER;
- Experiência com GIT;
- Experiência com ASP.NET MVC, WEB API, C# e Entity framework.
## Benefícios:
- Plano de saúde;
- Plano odontológico;
- Seguro de vida;
- Cartão de benefícios Swile (alimentação, auxílio internet, desconto em academia).
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (.NET/Javascript) na Revelar RH](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Categoria
Full-Stack
|
process
|
fullstack developer net javascript na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a revelar rh está em busca de fullstack developer para compor seu time responsabilidades desenvolvimento e acompanhamento de manutenção evolutivas e corretivas em sistemas desenvolvidos em net propor soluções e desenvolver novas tecnologias para melhoria dos sistemas manter a qualidade do código versionamento testes deploy e participação na priorização de funcionalidades horário de segunda à sexta das às revelar rh a revelar rh é uma empresa inovadora que busca alinhar os objetivos das empresas que têm carência de profissionais qualificados com os das pessoas que estão em busca de uma oportunidades no mercado de trabalho nosso trabalho visa o aprimoramento e racionalização de processos voltados à gestão de pessoas sendo realizado de forma personalizada para cada cliente atualmente atendemos várias empresas de tecnologia e softwares house nbsp habilidades asp net mvc c arquitetura de projetos mobile local belo horizonte minas gerais brazil requisitos superior completo ou em curso de análise e desenvolvimento de sistemas engenharia da computação ou áreas afins experiência e conhecimento correspondente ao perfil pleno sênior experiência com bootstrap experiência com sql server experiência com git experiência com asp net mvc web api c e entity framework benefícios plano de saúde plano odontológico seguro de vida cartão de benefícios swile alimentação auxílio internet desconto em academia como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado categoria full stack
| 1
|
8,918
| 4,351,323,509
|
IssuesEvent
|
2016-07-31 19:55:24
|
VOREStation/VOREStation
|
https://api.github.com/repos/VOREStation/VOREStation
|
closed
|
Never ending heartbeat sounds.
|
Pri: 2-Moderate Type: Bug Works in latest build
|
#### Brief description of the issue
The sound of a heart beating never ceases.
#### What you expected to happen
For it to stop.
#### What actually happened
It never stopped.
#### Steps to reproduce
1. Get gurgled.
2. Enjoy the annoying sound.
#### Additional info:
- **Server Revision**: Server revision: release - 2016-05-23
3bb482ee56a4ade2e02fbb46fa3746728c9a3b3e
- **Anything else you may wish to add** (Location if it's a mapping issue, etc)
|
1.0
|
Never ending heartbeat sounds. - #### Brief description of the issue
The sound of a heart beating never ceases.
#### What you expected to happen
For it to stop.
#### What actually happened
It never stopped.
#### Steps to reproduce
1. Get gurgled.
2. Enjoy the annoying sound.
#### Additional info:
- **Server Revision**: Server revision: release - 2016-05-23
3bb482ee56a4ade2e02fbb46fa3746728c9a3b3e
- **Anything else you may wish to add** (Location if it's a mapping issue, etc)
|
non_process
|
never ending heartbeat sounds brief description of the issue the sound of a heart beating never ceases what you expected to happen for it to stop what actually happened it never stopped steps to reproduce get gurgled enjoy the annoying sound additional info server revision server revision release anything else you may wish to add location if it s a mapping issue etc
| 0
|
19,195
| 25,321,324,410
|
IssuesEvent
|
2022-11-18 04:21:07
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Archlinux: "bazel crashed due to an internal error" when building media pipe
|
more data needed type: support / not a bug (process) team-OSS
|
### Description of the bug:
When attempting to build [mediapipe](https://github.com/google/mediapipe) on Archlinux, Bazel experiences an internal error.
Archlinux
bazel v5.3.2
java openjdk 19.0.1 2022-10-18
mediapipe v0.8.11
`
[user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion # requires Bazel 5.2.0 by default
[user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu
Starting local Bazel server and connecting to it...
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_absl' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_benchmark' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'flatbuffers' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'pybind11_bazel' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_googlesource_code_re2' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_protobuf' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_googletest' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_github_gflags_gflags' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'zlib' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'build_bazel_rules_apple' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'build_bazel_rules_swift' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'build_bazel_apple_support' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'xctestrunner' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'pybind11' because it already exists.
WARNING: /home/user/Git/mediapipe/mediapipe/framework/BUILD:54:24: in cc_library rule //mediapipe/framework:calculator_cc_proto: target '//mediapipe/framework:calculator_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use.
WARNING: /home/user/Git/mediapipe/mediapipe/framework/tool/BUILD:185:24: in cc_library rule //mediapipe/framework/tool:field_data_cc_proto: target '//mediapipe/framework/tool:field_data_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use.
INFO: Analyzed target //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu (140 packages loaded, 5537 targets configured).
INFO: Found 1 target...
[0 / 8] 3 actions, 0 running
[Prepa] BazelWorkspaceStatusAction stable-status.txt
[Prepa] Writing file mediapipe/examples/desktop/hand_tracking/hand_tracking_cpu-2.params
[Prepa] Creating source manifest for //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu
FATAL: bazel crashed due to an internal error. Printing stack trace:
java.lang.ExceptionInInitializerError
at com.google.devtools.build.lib.actions.ParameterFile.writeContent(ParameterFile.java:118)
at com.google.devtools.build.lib.actions.ParameterFile.writeParameterFile(ParameterFile.java:111)
at com.google.devtools.build.lib.analysis.actions.ParameterFileWriteAction$ParamFileWriter.writeOutputFile(ParameterFileWriteAction.java:170)
at com.google.devtools.build.lib.exec.FileWriteStrategy.beginWriteOutputToFile(FileWriteStrategy.java:58)
at com.google.devtools.build.lib.analysis.actions.FileWriteActionContext.beginWriteOutputToFile(FileWriteActionContext.java:49)
at com.google.devtools.build.lib.analysis.actions.AbstractFileWriteAction.beginExecution(AbstractFileWriteAction.java:66)
at com.google.devtools.build.lib.actions.Action.execute(Action.java:133)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169)
at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590)
at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)
at java.base/java.lang.Thread.run(Thread.java:1589)
Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make java.lang.String(byte[],byte) accessible: module java.base does not "opens java.lang" to unnamed module @7ed7259e
at java.base/java.lang.reflect.AccessibleObject.throwInaccessibleObjectException(AccessibleObject.java:387)
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:363)
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:311)
at java.base/java.lang.reflect.Constructor.checkCanSetAccessible(Constructor.java:192)
at java.base/java.lang.reflect.Constructor.setAccessible(Constructor.java:185)
at com.google.devtools.build.lib.unsafe.StringUnsafe.<init>(StringUnsafe.java:75)
at com.google.devtools.build.lib.unsafe.StringUnsafe.initInstance(StringUnsafe.java:56)
at com.google.devtools.build.lib.unsafe.StringUnsafe.<clinit>(StringUnsafe.java:37)
... 21 more
`
Thanks!
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
[user@system mediapipe]$ git clone https://github.com/google/mediapipe.git
[user@system mediapipe]$ cd mediapipe
[user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion
[user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop
### Which operating system are you running Bazel on?
Archlinux
### What is the output of `bazel info release`?
release 5.3.2
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
No.
### Any other information, logs, or outputs that you want to share?
See description.
|
1.0
|
Archlinux: "bazel crashed due to an internal error" when building media pipe - ### Description of the bug:
When attempting to build [mediapipe](https://github.com/google/mediapipe) on Archlinux, Bazel experiences an internal error.
Archlinux
bazel v5.3.2
java openjdk 19.0.1 2022-10-18
mediapipe v0.8.11
`
[user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion # requires Bazel 5.2.0 by default
[user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu
Starting local Bazel server and connecting to it...
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_absl' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_benchmark' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'flatbuffers' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'pybind11_bazel' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_googlesource_code_re2' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_protobuf' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_google_googletest' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'com_github_gflags_gflags' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'zlib' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'build_bazel_rules_apple' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'build_bazel_rules_swift' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'build_bazel_apple_support' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'xctestrunner' because it already exists.
DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14:
Warning: skipping import of repository 'pybind11' because it already exists.
WARNING: /home/user/Git/mediapipe/mediapipe/framework/BUILD:54:24: in cc_library rule //mediapipe/framework:calculator_cc_proto: target '//mediapipe/framework:calculator_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use.
WARNING: /home/user/Git/mediapipe/mediapipe/framework/tool/BUILD:185:24: in cc_library rule //mediapipe/framework/tool:field_data_cc_proto: target '//mediapipe/framework/tool:field_data_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use.
INFO: Analyzed target //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu (140 packages loaded, 5537 targets configured).
INFO: Found 1 target...
[0 / 8] 3 actions, 0 running
[Prepa] BazelWorkspaceStatusAction stable-status.txt
[Prepa] Writing file mediapipe/examples/desktop/hand_tracking/hand_tracking_cpu-2.params
[Prepa] Creating source manifest for //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu
FATAL: bazel crashed due to an internal error. Printing stack trace:
java.lang.ExceptionInInitializerError
at com.google.devtools.build.lib.actions.ParameterFile.writeContent(ParameterFile.java:118)
at com.google.devtools.build.lib.actions.ParameterFile.writeParameterFile(ParameterFile.java:111)
at com.google.devtools.build.lib.analysis.actions.ParameterFileWriteAction$ParamFileWriter.writeOutputFile(ParameterFileWriteAction.java:170)
at com.google.devtools.build.lib.exec.FileWriteStrategy.beginWriteOutputToFile(FileWriteStrategy.java:58)
at com.google.devtools.build.lib.analysis.actions.FileWriteActionContext.beginWriteOutputToFile(FileWriteActionContext.java:49)
at com.google.devtools.build.lib.analysis.actions.AbstractFileWriteAction.beginExecution(AbstractFileWriteAction.java:66)
at com.google.devtools.build.lib.actions.Action.execute(Action.java:133)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152)
at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91)
at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349)
at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169)
at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590)
at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)
at java.base/java.lang.Thread.run(Thread.java:1589)
Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make java.lang.String(byte[],byte) accessible: module java.base does not "opens java.lang" to unnamed module @7ed7259e
at java.base/java.lang.reflect.AccessibleObject.throwInaccessibleObjectException(AccessibleObject.java:387)
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:363)
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:311)
at java.base/java.lang.reflect.Constructor.checkCanSetAccessible(Constructor.java:192)
at java.base/java.lang.reflect.Constructor.setAccessible(Constructor.java:185)
at com.google.devtools.build.lib.unsafe.StringUnsafe.<init>(StringUnsafe.java:75)
at com.google.devtools.build.lib.unsafe.StringUnsafe.initInstance(StringUnsafe.java:56)
at com.google.devtools.build.lib.unsafe.StringUnsafe.<clinit>(StringUnsafe.java:37)
... 21 more
`
Thanks!
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
[user@system mediapipe]$ git clone https://github.com/google/mediapipe.git
[user@system mediapipe]$ cd mediapipe
[user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion
[user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop
### Which operating system are you running Bazel on?
Archlinux
### What is the output of `bazel info release`?
release 5.3.2
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
No.
### Any other information, logs, or outputs that you want to share?
See description.
|
process
|
archlinux bazel crashed due to an internal error when building media pipe description of the bug when attempting to build on archlinux bazel experiences an internal error archlinux bazel java openjdk mediapipe bazel version sed s bazel bazelversion requires bazel by default bazel build c opt define mediapipe disable gpu mediapipe examples desktop hand tracking hand tracking cpu starting local bazel server and connecting to it debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google absl because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google benchmark because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository flatbuffers because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository bazel because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com googlesource code because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google protobuf because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google googletest because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com github gflags gflags because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository zlib because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository build bazel rules apple because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository build bazel rules swift because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository build bazel apple support because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository xctestrunner because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository because it already exists warning home user git mediapipe mediapipe framework build in cc library rule mediapipe framework calculator cc proto target mediapipe framework calculator cc proto depends on deprecated target com google protobuf cc wkt protos only for backward compatibility do not use warning home user git mediapipe mediapipe framework tool build in cc library rule mediapipe framework tool field data cc proto target mediapipe framework tool field data cc proto depends on deprecated target com google protobuf cc wkt protos only for backward compatibility do not use info analyzed target mediapipe examples desktop hand tracking hand tracking cpu packages loaded targets configured info found target actions running bazelworkspacestatusaction stable status txt writing file mediapipe examples desktop hand tracking hand tracking cpu params creating source manifest for mediapipe examples desktop hand tracking hand tracking cpu fatal bazel crashed due to an internal error printing stack trace java lang exceptionininitializererror at com google devtools build lib actions parameterfile writecontent parameterfile java at com google devtools build lib actions parameterfile writeparameterfile parameterfile java at com google devtools build lib analysis actions parameterfilewriteaction paramfilewriter writeoutputfile parameterfilewriteaction java at com google devtools build lib exec filewritestrategy beginwriteoutputtofile filewritestrategy java at com google devtools build lib analysis actions filewriteactioncontext beginwriteoutputtofile filewriteactioncontext java at com google devtools build lib analysis actions abstractfilewriteaction beginexecution abstractfilewriteaction java at com google devtools build lib actions action execute action java at com google devtools build lib skyframe skyframeactionexecutor execute skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner continueaction skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner run skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionstate runstatemachine actionexecutionstate java at com google devtools build lib skyframe actionexecutionstate getresultordependonfuture actionexecutionstate java at com google devtools build lib skyframe skyframeactionexecutor executeaction skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionfunction checkcacheandexecuteifneeded actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction computeinternal actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction compute actionexecutionfunction java at com google devtools build skyframe abstractparallelevaluator evaluate run abstractparallelevaluator java at com google devtools build lib concurrent abstractqueuevisitor wrappedrunnable run abstractqueuevisitor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java lang reflect inaccessibleobjectexception unable to make java lang string byte byte accessible module java base does not opens java lang to unnamed module at java base java lang reflect accessibleobject throwinaccessibleobjectexception accessibleobject java at java base java lang reflect accessibleobject checkcansetaccessible accessibleobject java at java base java lang reflect accessibleobject checkcansetaccessible accessibleobject java at java base java lang reflect constructor checkcansetaccessible constructor java at java base java lang reflect constructor setaccessible constructor java at com google devtools build lib unsafe stringunsafe stringunsafe java at com google devtools build lib unsafe stringunsafe initinstance stringunsafe java at com google devtools build lib unsafe stringunsafe stringunsafe java more thanks what s the simplest easiest way to reproduce this bug please provide a minimal example if possible git clone cd mediapipe bazel version sed s bazel bazelversion bazel build c opt define mediapipe disable gpu mediapipe examples desktop which operating system are you running bazel on archlinux what is the output of bazel info release release if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no any other information logs or outputs that you want to share see description
| 1
|
73,001
| 3,398,568,619
|
IssuesEvent
|
2015-12-02 05:01:39
|
tomreece/pucauto
|
https://api.github.com/repos/tomreece/pucauto
|
closed
|
Make easier to run for Windows users
|
high priority large
|
@droogans How do Windows users typically install dependencies that us Mac users would install with pip? I want them to be able to use my `requirements.txt` file not have to manually go grab every dependency I keep adding. I'm about to merge a PR that adds Six to the project. Thoughts?
|
1.0
|
Make easier to run for Windows users - @droogans How do Windows users typically install dependencies that us Mac users would install with pip? I want them to be able to use my `requirements.txt` file not have to manually go grab every dependency I keep adding. I'm about to merge a PR that adds Six to the project. Thoughts?
|
non_process
|
make easier to run for windows users droogans how do windows users typically install dependencies that us mac users would install with pip i want them to be able to use my requirements txt file not have to manually go grab every dependency i keep adding i m about to merge a pr that adds six to the project thoughts
| 0
|
22,106
| 30,635,987,384
|
IssuesEvent
|
2023-07-24 17:48:16
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
roblox-pyc 1.24.102 has 5 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.24.102",
"result": {
"issues": 5,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:143",
"code": " subprocess.call([\"npm\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:149",
"code": " subprocess.call([\"rbxtsc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:188",
"code": " subprocess.call([\"wally\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:198",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:205",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpntaz4g58/roblox-pyc"
}
}```
|
1.0
|
roblox-pyc 1.24.102 has 5 GuardDog issues - https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.24.102",
"result": {
"issues": 5,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:143",
"code": " subprocess.call([\"npm\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:149",
"code": " subprocess.call([\"rbxtsc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:188",
"code": " subprocess.call([\"wally\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:198",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:205",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpntaz4g58/roblox-pyc"
}
}```
|
process
|
roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp roblox pyc
| 1
|
238,574
| 26,135,891,805
|
IssuesEvent
|
2022-12-29 12:04:39
|
MatBenfield/news
|
https://api.github.com/repos/MatBenfield/news
|
closed
|
[SecurityWeek] EarSpy: Spying on Phone Calls via Ear Speaker Vibrations Captured by Accelerometer
|
SecurityWeek Stale
|
**As smartphone manufacturers are improving the ear speakers in their devices, it can become easier for malicious actors to leverage a particular side-channel for eavesdropping on a targeted user’s conversations, according to a team of researchers from several universities in the United States.**
[read more](https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer)
<https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer>
|
True
|
[SecurityWeek] EarSpy: Spying on Phone Calls via Ear Speaker Vibrations Captured by Accelerometer -
**As smartphone manufacturers are improving the ear speakers in their devices, it can become easier for malicious actors to leverage a particular side-channel for eavesdropping on a targeted user’s conversations, according to a team of researchers from several universities in the United States.**
[read more](https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer)
<https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer>
|
non_process
|
earspy spying on phone calls via ear speaker vibrations captured by accelerometer as smartphone manufacturers are improving the ear speakers in their devices it can become easier for malicious actors to leverage a particular side channel for eavesdropping on a targeted user’s conversations according to a team of researchers from several universities in the united states
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.