Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
21,367
29,194,080,693
IssuesEvent
2023-05-20 00:31:50
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] QA Analyst na Coodesh
SALVADOR TESTE PJ JAVASCRIPT RUBY REQUISITOS SELENIUM CUCUMBER REMOTO CYPRESS PROCESSOS GITHUB UMA QUALIDADE VENDAS QA TESTES AUTOMATIZADOS METODOLOGIAS ÁGEIS RSPEC AUTOMAÇÃO DE TESTES Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>BaladAPP </strong>está em busca de <strong><ins>QA Analyst</ins></strong> para compor seu time!<br><br>A BaladAPP é uma das maiores ticketeiras do Brasil, atuando de norte a sul, atendendo mais de 200 mil clientes mensalmente. <br>Buscamos sempre trazer soluções para produtores de eventos auxiliando desde o início das vendas online até o dia do evento. <br><br><strong>Responsabilidades:</strong></p> <ul> <li>Estamos procurando por um profissional de qualidade de software para se juntar ao nosso time e estruturar nosso setor de qualidade. Como membro da equipe, você será responsável por testar nossos sistemas, fazer a gestão das nossas versões e manter nosso changelog. Além disso, você terá a oportunidade de evoluir e avaliar nossos testes automatizados, a fim de tornar nossos processos de teste mais eficientes.</li> </ul> <p><br><br></p> ## BALADAPP: <p>A BaladAPP é uma empresa especializada em vendas de ingressos para eventos nacionais e internacionais, focados em shows e espetáculos, funcionando como marketplace, trabalhamos com eventos de pequeno à grande porte.</p> </p> ## Habilidades: - RSpec - Cucumber - Cypress - Selenium - Automação de Testes ## Local: 100% Remoto ## Requisitos: - Experiência anterior como Analista de Testes/QA; - Experiência com testes automatizados com uma destas ferramentas Selenium, Cypress, Cucumber (Preferencialmente com RSpec); - Conhecimento em Ruby; - Conhecimentos em processos e metodologias ágeis; - Habilidades de resolução de problemas e pensamento crítico. ## Diferenciais: - Conhecimento de linguagens de programação Javascript. ## Benefícios: - Horários flexíveis; - 30 dias de férias remuneradas no PJ; - Plano de Saúde; - Plano Odontológico. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [QA Analyst na BALADAPP](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Testes/Q.A
1.0
[Remoto] QA Analyst na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>BaladAPP </strong>está em busca de <strong><ins>QA Analyst</ins></strong> para compor seu time!<br><br>A BaladAPP é uma das maiores ticketeiras do Brasil, atuando de norte a sul, atendendo mais de 200 mil clientes mensalmente. <br>Buscamos sempre trazer soluções para produtores de eventos auxiliando desde o início das vendas online até o dia do evento. <br><br><strong>Responsabilidades:</strong></p> <ul> <li>Estamos procurando por um profissional de qualidade de software para se juntar ao nosso time e estruturar nosso setor de qualidade. Como membro da equipe, você será responsável por testar nossos sistemas, fazer a gestão das nossas versões e manter nosso changelog. Além disso, você terá a oportunidade de evoluir e avaliar nossos testes automatizados, a fim de tornar nossos processos de teste mais eficientes.</li> </ul> <p><br><br></p> ## BALADAPP: <p>A BaladAPP é uma empresa especializada em vendas de ingressos para eventos nacionais e internacionais, focados em shows e espetáculos, funcionando como marketplace, trabalhamos com eventos de pequeno à grande porte.</p> </p> ## Habilidades: - RSpec - Cucumber - Cypress - Selenium - Automação de Testes ## Local: 100% Remoto ## Requisitos: - Experiência anterior como Analista de Testes/QA; - Experiência com testes automatizados com uma destas ferramentas Selenium, Cypress, Cucumber (Preferencialmente com RSpec); - Conhecimento em Ruby; - Conhecimentos em processos e metodologias ágeis; - Habilidades de resolução de problemas e pensamento crítico. ## Diferenciais: - Conhecimento de linguagens de programação Javascript. ## Benefícios: - Horários flexíveis; - 30 dias de férias remuneradas no PJ; - Plano de Saúde; - Plano Odontológico. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [QA Analyst na BALADAPP](https://coodesh.com/vagas/qa-analyst-202942193?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Testes/Q.A
process
qa analyst na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a baladapp está em busca de qa analyst para compor seu time a baladapp é uma das maiores ticketeiras do brasil atuando de norte a sul atendendo mais de mil clientes mensalmente buscamos sempre trazer soluções para produtores de eventos auxiliando desde o início das vendas online até o dia do evento responsabilidades estamos procurando por um profissional de qualidade de software para se juntar ao nosso time e estruturar nosso setor de qualidade como membro da equipe você será responsável por testar nossos sistemas fazer a gestão das nossas versões e manter nosso changelog além disso você terá a oportunidade de evoluir e avaliar nossos testes automatizados a fim de tornar nossos processos de teste mais eficientes baladapp a baladapp é uma empresa especializada em vendas de ingressos para eventos nacionais e internacionais focados em shows e espetáculos funcionando como marketplace trabalhamos com eventos de pequeno à grande porte habilidades rspec cucumber cypress selenium automação de testes local remoto requisitos experiência anterior como analista de testes qa experiência com testes automatizados com uma destas ferramentas selenium cypress cucumber preferencialmente com rspec conhecimento em ruby conhecimentos em processos e metodologias ágeis habilidades de resolução de problemas e pensamento crítico diferenciais conhecimento de linguagens de programação javascript benefícios horários flexíveis dias de férias remuneradas no pj plano de saúde plano odontológico como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria testes q a
1
7,259
10,420,471,777
IssuesEvent
2019-09-16 00:42:00
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
In place processing multipart to singleparts does not handle unique constraints
Bug Processing
Provider errors are generated due to failure of UNIQUE constraint when you try to save after running Multipart to singleparts in a geopackage layer. **How to Reproduce** 1. Download and open a copy of ‘Multi_to_Singleparts_FID_Bug.qgz’ (zip attached below) 2. Open the Attribute Table for the ‘Multi_to_Singleparts_FID_Bug’ layer. 3. Select the multipart feature with FID value of 4. 4. Ctrl + K 5. “ef multipart to singleparts” 6. Observe there are now 5 features selected in the attribute table with FID value of 4 (This is a bug). 7. Undo, and the edit is successfully reversed. 8. Repeat steps 3 to 6. 9. Click ‘Save Layer Edits’ 10. Observe there are 6 Commit errors, including: Could not commit changes to layer Multi_to_Singleparts_FID_Bug Errors: SUCCESS: 1 feature(s) deleted. ERROR: 5 feature(s) not added. Provider errors: OGR error creating feature -8: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid OGR error creating feature -9: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid OGR error creating feature -10: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid OGR error creating feature -11: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid 11. On the Map Canvas, marquee select the feature(s) at the south-west corner. 12. In the attribute table, observe that the title bar reports 2 features selected, but there is only one feature selected in the table (bug in attribute table failing to update). 13. Close and re-open the attribute table, and observe there are now 2 features selected out of 6 features with FID value of 4. 14. Either ‘Undo’ or ‘Toggle Edit & Discard Changes’. 15. Observe that 5 of the 6 FID value-4 features have been deleted, leaving only one remaining. Workaround: 16. Repeat steps 1 to 6 17. In the attribute table, with the relevant features selected, change the FID value to null. 18. Save Layer Edits. 19. Observe that unique values have now been autogenerated. **QGIS and OS versions** QGIS 3.4.11 (LTR), 3.8.2, and 3.9 (nightly build) [Multi_to_Singleparts_FID_Bug.zip](https://github.com/qgis/QGIS/files/3590496/Multi_to_Singleparts_FID_Bug.zip)
1.0
In place processing multipart to singleparts does not handle unique constraints - Provider errors are generated due to failure of UNIQUE constraint when you try to save after running Multipart to singleparts in a geopackage layer. **How to Reproduce** 1. Download and open a copy of ‘Multi_to_Singleparts_FID_Bug.qgz’ (zip attached below) 2. Open the Attribute Table for the ‘Multi_to_Singleparts_FID_Bug’ layer. 3. Select the multipart feature with FID value of 4. 4. Ctrl + K 5. “ef multipart to singleparts” 6. Observe there are now 5 features selected in the attribute table with FID value of 4 (This is a bug). 7. Undo, and the edit is successfully reversed. 8. Repeat steps 3 to 6. 9. Click ‘Save Layer Edits’ 10. Observe there are 6 Commit errors, including: Could not commit changes to layer Multi_to_Singleparts_FID_Bug Errors: SUCCESS: 1 feature(s) deleted. ERROR: 5 feature(s) not added. Provider errors: OGR error creating feature -8: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid OGR error creating feature -9: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid OGR error creating feature -10: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid OGR error creating feature -11: failed to execute insert : UNIQUE constraint failed: Multi_to_Singleparts_FID_Bug.fid 11. On the Map Canvas, marquee select the feature(s) at the south-west corner. 12. In the attribute table, observe that the title bar reports 2 features selected, but there is only one feature selected in the table (bug in attribute table failing to update). 13. Close and re-open the attribute table, and observe there are now 2 features selected out of 6 features with FID value of 4. 14. Either ‘Undo’ or ‘Toggle Edit & Discard Changes’. 15. Observe that 5 of the 6 FID value-4 features have been deleted, leaving only one remaining. Workaround: 16. Repeat steps 1 to 6 17. In the attribute table, with the relevant features selected, change the FID value to null. 18. Save Layer Edits. 19. Observe that unique values have now been autogenerated. **QGIS and OS versions** QGIS 3.4.11 (LTR), 3.8.2, and 3.9 (nightly build) [Multi_to_Singleparts_FID_Bug.zip](https://github.com/qgis/QGIS/files/3590496/Multi_to_Singleparts_FID_Bug.zip)
process
in place processing multipart to singleparts does not handle unique constraints provider errors are generated due to failure of unique constraint when you try to save after running multipart to singleparts in a geopackage layer how to reproduce download and open a copy of ‘multi to singleparts fid bug qgz’ zip attached below open the attribute table for the ‘multi to singleparts fid bug’ layer select the multipart feature with fid value of ctrl k “ef multipart to singleparts” observe there are now features selected in the attribute table with fid value of this is a bug undo and the edit is successfully reversed repeat steps to click ‘save layer edits’ observe there are commit errors including could not commit changes to layer multi to singleparts fid bug errors success feature s deleted error feature s not added provider errors ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid ogr error creating feature failed to execute insert unique constraint failed multi to singleparts fid bug fid on the map canvas marquee select the feature s at the south west corner in the attribute table observe that the title bar reports features selected but there is only one feature selected in the table bug in attribute table failing to update close and re open the attribute table and observe there are now features selected out of features with fid value of either ‘undo’ or ‘toggle edit discard changes’ observe that of the fid value features have been deleted leaving only one remaining workaround repeat steps to in the attribute table with the relevant features selected change the fid value to null save layer edits observe that unique values have now been autogenerated qgis and os versions qgis ltr and nightly build
1
71,721
3,367,617,921
IssuesEvent
2015-11-22 10:19:05
music-encoding/music-encoding
https://api.github.com/repos/music-encoding/music-encoding
closed
The element head is not allowed in projectDesc
Priority: Medium
_From [siggelun...@gmail.com](https://code.google.com/u/110461478002540803867/) on December 19, 2013 03:06:13_ What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? What version of the product are you using? On what operating system? Please provide any additional information below. In my view \<projectDesc> and elements with similar function should have a more fully fledge support for prose. _Original issue: http://code.google.com/p/music-encoding/issues/detail?id=187_
1.0
The element head is not allowed in projectDesc - _From [siggelun...@gmail.com](https://code.google.com/u/110461478002540803867/) on December 19, 2013 03:06:13_ What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? What version of the product are you using? On what operating system? Please provide any additional information below. In my view \<projectDesc> and elements with similar function should have a more fully fledge support for prose. _Original issue: http://code.google.com/p/music-encoding/issues/detail?id=187_
non_process
the element head is not allowed in projectdesc from on december what steps will reproduce the problem what is the expected output what do you see instead what version of the product are you using on what operating system please provide any additional information below in my view and elements with similar function should have a more fully fledge support for prose original issue
0
17,141
22,686,040,208
IssuesEvent
2022-07-04 14:11:13
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Enhance randomized process tests with start process instance anywhere ability
team/process-automation
To comprehensively test the start process instance anywhere feature, we'll need to enhance the randomized process tests with the ability to start the process instance anywhere. Blocked by #9390 #9391 ## Out of scope - starting the process instance inside a multi-instance - starting the process instance inside a call activity
1.0
Enhance randomized process tests with start process instance anywhere ability - To comprehensively test the start process instance anywhere feature, we'll need to enhance the randomized process tests with the ability to start the process instance anywhere. Blocked by #9390 #9391 ## Out of scope - starting the process instance inside a multi-instance - starting the process instance inside a call activity
process
enhance randomized process tests with start process instance anywhere ability to comprehensively test the start process instance anywhere feature we ll need to enhance the randomized process tests with the ability to start the process instance anywhere blocked by out of scope starting the process instance inside a multi instance starting the process instance inside a call activity
1
7,406
10,525,900,731
IssuesEvent
2019-09-30 15:56:15
Python-Markdown/markdown
https://api.github.com/repos/Python-Markdown/markdown
reopened
Python Version Support Timeline
process
Just noting this so I can find it later. Python defines the status of its versions [here][1]. The following end-of life cycles are currently scheduled: * <del>Python 3.4 2019-03-16</del> * Python 2.7 2020-01-01 * Python 3.5 2020-09-13 * Python 3.6 2021-12-23 * Python 3.7 2023-06-27 * Python 3.8 2024-10 Also of interest is https://python3statement.org/ where various projects have signed on to a pledge to discontinue support for Python 2,7 on or before 2020. I'm indifferent to joining the pledge, but the arguments given for dropping support could be useful when we take that move. I intend to keep this issue open indefinitely to track future updates. As a general rule, I expect we will drop support for a Python version with our next minor release after Python has dropped support. That said, if/when Travis drops support, our tests stop passing, so we have occasionally dropped support quickly in the past. Of course, as per our Contributing Guidelines, such a change should not happen in a bugfix release. [1]: https://devguide.python.org/#status-of-python-branches
1.0
Python Version Support Timeline - Just noting this so I can find it later. Python defines the status of its versions [here][1]. The following end-of life cycles are currently scheduled: * <del>Python 3.4 2019-03-16</del> * Python 2.7 2020-01-01 * Python 3.5 2020-09-13 * Python 3.6 2021-12-23 * Python 3.7 2023-06-27 * Python 3.8 2024-10 Also of interest is https://python3statement.org/ where various projects have signed on to a pledge to discontinue support for Python 2,7 on or before 2020. I'm indifferent to joining the pledge, but the arguments given for dropping support could be useful when we take that move. I intend to keep this issue open indefinitely to track future updates. As a general rule, I expect we will drop support for a Python version with our next minor release after Python has dropped support. That said, if/when Travis drops support, our tests stop passing, so we have occasionally dropped support quickly in the past. Of course, as per our Contributing Guidelines, such a change should not happen in a bugfix release. [1]: https://devguide.python.org/#status-of-python-branches
process
python version support timeline just noting this so i can find it later python defines the status of its versions the following end of life cycles are currently scheduled python python python python python python also of interest is where various projects have signed on to a pledge to discontinue support for python on or before i m indifferent to joining the pledge but the arguments given for dropping support could be useful when we take that move i intend to keep this issue open indefinitely to track future updates as a general rule i expect we will drop support for a python version with our next minor release after python has dropped support that said if when travis drops support our tests stop passing so we have occasionally dropped support quickly in the past of course as per our contributing guidelines such a change should not happen in a bugfix release
1
18,085
24,107,887,533
IssuesEvent
2022-09-20 08:58:04
googleapis/dotnet-spanner-nhibernate
https://api.github.com/repos/googleapis/dotnet-spanner-nhibernate
closed
Dependency Dashboard
type: process api: spanner
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/benchmarkdotnet-0.x -->[chore(deps): update dependency benchmarkdotnet to v0.13.2](../pull/98) - [ ] <!-- rebase-branch=renovate/docker.dotnet-3.x -->[chore(deps): update dependency docker.dotnet to v3.125.12](../pull/102) - [ ] <!-- rebase-branch=renovate/gcp-releasetool-1.x -->[chore(deps): update dependency gcp-releasetool to v1.8.8](../pull/109) - [ ] <!-- rebase-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2.11.1](../pull/112) - [ ] <!-- rebase-branch=renovate/keyring-23.x -->[chore(deps): update dependency keyring to v23.9.3](../pull/111) - [ ] <!-- rebase-branch=renovate/vstest-monorepo -->[chore(deps): update dependency microsoft.net.test.sdk to v17.3.1](../pull/99) - [ ] <!-- rebase-branch=renovate/nhibernate-5.x -->[chore(deps): update dependency nhibernate to v5.3.13](../pull/101) - [ ] <!-- rebase-branch=renovate/pyjwt-2.x -->[chore(deps): update dependency pyjwt to v2.5.0](../pull/110) - [ ] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once** ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/certifi-2022.x -->[chore(deps): update dependency certifi to v2022.9.14](../pull/106) - [ ] <!-- recreate-branch=renovate/click-8.x -->[chore(deps): update dependency click to v8.1.3](../pull/104) - [ ] <!-- recreate-branch=renovate/google.cloud.spanner.data-4.x -->[chore(deps): update dependency google.cloud.spanner.data to v4](../pull/94) - [ ] <!-- recreate-branch=renovate/protobuf-4.x -->[chore(deps): update dependency protobuf to v4](../pull/105) ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/build-and-test-on-windows.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> <details><summary>.github/workflows/build-and-test.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> <details><summary>.github/workflows/integration-tests-on-emulator.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> <details><summary>.github/workflows/integration-tests-on-production.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> </blockquote> </details> <details><summary>nuget</summary> <blockquote> <details><summary>Google.Cloud.Spanner.Connection.Tests/Google.Cloud.Spanner.Connection.Tests.csproj</summary> - `coverlet.collector 3.1.2` - `xunit.runner.visualstudio 2.4.5` - `Xunit.Combinatorial 1.4.1` - `xunit 2.4.2` - `Microsoft.NET.Test.Sdk 17.3.0` </details> <details><summary>Google.Cloud.Spanner.Connection/Google.Cloud.Spanner.Connection.csproj</summary> - `Google.Cloud.Spanner.Data 3.15.1` </details> <details><summary>Google.Cloud.Spanner.NHibernate.Benchmarks/Google.Cloud.Spanner.NHibernate.Benchmarks.csproj</summary> - `BenchmarkDotNet 0.13.1` </details> <details><summary>Google.Cloud.Spanner.NHibernate.IntegrationTests/Google.Cloud.Spanner.NHibernate.IntegrationTests.csproj</summary> - `Xunit.SkippableFact 1.4.13` - `coverlet.collector 3.1.2` - `xunit.runner.visualstudio 2.4.5` - `Xunit.Combinatorial 1.4.1` - `xunit 2.4.2` - `Microsoft.NET.Test.Sdk 17.3.0` </details> <details><summary>Google.Cloud.Spanner.NHibernate.Samples/Google.Cloud.Spanner.NHibernate.Samples.csproj</summary> - `Docker.DotNet 3.125.10` </details> <details><summary>Google.Cloud.Spanner.NHibernate.Tests/Google.Cloud.Spanner.NHibernate.Tests.csproj</summary> - `coverlet.collector 3.1.2` - `xunit.runner.visualstudio 2.4.5` - `Xunit.Combinatorial 1.4.1` - `xunit 2.4.2` - `Microsoft.NET.Test.Sdk 17.3.0` </details> <details><summary>Google.Cloud.Spanner.NHibernate/Google.Cloud.Spanner.NHibernate.csproj</summary> - `NHibernate 5.3.12` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>.kokoro/requirements.txt</summary> - `attrs ==22.1.0` - `cachetools ==5.2.0` - `certifi ==2022.6.15.2` - `cffi ==1.15.1` - `charset-normalizer ==2.1.1` - `click ==8.0.4` - `colorama ==0.4.5` - `cryptography ==38.0.1` - `gcp-releasetool ==1.8.7` - `google-auth ==2.11.0` - `idna ==3.4` - `importlib-metadata ==4.12.0` - `jaraco-classes ==3.2.2` - `jinja2 ==3.1.2` - `keyring ==23.9.1` - `markupsafe ==2.1.1` - `more-itertools ==8.14.0` - `packaging ==21.3` - `protobuf ==3.20.2` - `pyasn1 ==0.4.8` - `pyasn1-modules ==0.2.8` - `pycparser ==2.21` - `pyjwt ==2.4.0` - `pyparsing ==3.0.9` - `pyperclip ==1.8.2` - `python-dateutil ==2.8.2` - `pywin32-ctypes ==0.2.0` - `requests ==2.28.1` - `rsa ==4.9` - `six ==1.16.0` - `typing-extensions ==4.3.0` - `urllib3 ==1.26.12` - `zipp ==3.8.1` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/benchmarkdotnet-0.x -->[chore(deps): update dependency benchmarkdotnet to v0.13.2](../pull/98) - [ ] <!-- rebase-branch=renovate/docker.dotnet-3.x -->[chore(deps): update dependency docker.dotnet to v3.125.12](../pull/102) - [ ] <!-- rebase-branch=renovate/gcp-releasetool-1.x -->[chore(deps): update dependency gcp-releasetool to v1.8.8](../pull/109) - [ ] <!-- rebase-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2.11.1](../pull/112) - [ ] <!-- rebase-branch=renovate/keyring-23.x -->[chore(deps): update dependency keyring to v23.9.3](../pull/111) - [ ] <!-- rebase-branch=renovate/vstest-monorepo -->[chore(deps): update dependency microsoft.net.test.sdk to v17.3.1](../pull/99) - [ ] <!-- rebase-branch=renovate/nhibernate-5.x -->[chore(deps): update dependency nhibernate to v5.3.13](../pull/101) - [ ] <!-- rebase-branch=renovate/pyjwt-2.x -->[chore(deps): update dependency pyjwt to v2.5.0](../pull/110) - [ ] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once** ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/certifi-2022.x -->[chore(deps): update dependency certifi to v2022.9.14](../pull/106) - [ ] <!-- recreate-branch=renovate/click-8.x -->[chore(deps): update dependency click to v8.1.3](../pull/104) - [ ] <!-- recreate-branch=renovate/google.cloud.spanner.data-4.x -->[chore(deps): update dependency google.cloud.spanner.data to v4](../pull/94) - [ ] <!-- recreate-branch=renovate/protobuf-4.x -->[chore(deps): update dependency protobuf to v4](../pull/105) ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/build-and-test-on-windows.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> <details><summary>.github/workflows/build-and-test.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> <details><summary>.github/workflows/integration-tests-on-emulator.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> <details><summary>.github/workflows/integration-tests-on-production.yml</summary> - `actions/checkout v3` - `actions/setup-dotnet v2` </details> </blockquote> </details> <details><summary>nuget</summary> <blockquote> <details><summary>Google.Cloud.Spanner.Connection.Tests/Google.Cloud.Spanner.Connection.Tests.csproj</summary> - `coverlet.collector 3.1.2` - `xunit.runner.visualstudio 2.4.5` - `Xunit.Combinatorial 1.4.1` - `xunit 2.4.2` - `Microsoft.NET.Test.Sdk 17.3.0` </details> <details><summary>Google.Cloud.Spanner.Connection/Google.Cloud.Spanner.Connection.csproj</summary> - `Google.Cloud.Spanner.Data 3.15.1` </details> <details><summary>Google.Cloud.Spanner.NHibernate.Benchmarks/Google.Cloud.Spanner.NHibernate.Benchmarks.csproj</summary> - `BenchmarkDotNet 0.13.1` </details> <details><summary>Google.Cloud.Spanner.NHibernate.IntegrationTests/Google.Cloud.Spanner.NHibernate.IntegrationTests.csproj</summary> - `Xunit.SkippableFact 1.4.13` - `coverlet.collector 3.1.2` - `xunit.runner.visualstudio 2.4.5` - `Xunit.Combinatorial 1.4.1` - `xunit 2.4.2` - `Microsoft.NET.Test.Sdk 17.3.0` </details> <details><summary>Google.Cloud.Spanner.NHibernate.Samples/Google.Cloud.Spanner.NHibernate.Samples.csproj</summary> - `Docker.DotNet 3.125.10` </details> <details><summary>Google.Cloud.Spanner.NHibernate.Tests/Google.Cloud.Spanner.NHibernate.Tests.csproj</summary> - `coverlet.collector 3.1.2` - `xunit.runner.visualstudio 2.4.5` - `Xunit.Combinatorial 1.4.1` - `xunit 2.4.2` - `Microsoft.NET.Test.Sdk 17.3.0` </details> <details><summary>Google.Cloud.Spanner.NHibernate/Google.Cloud.Spanner.NHibernate.csproj</summary> - `NHibernate 5.3.12` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>.kokoro/requirements.txt</summary> - `attrs ==22.1.0` - `cachetools ==5.2.0` - `certifi ==2022.6.15.2` - `cffi ==1.15.1` - `charset-normalizer ==2.1.1` - `click ==8.0.4` - `colorama ==0.4.5` - `cryptography ==38.0.1` - `gcp-releasetool ==1.8.7` - `google-auth ==2.11.0` - `idna ==3.4` - `importlib-metadata ==4.12.0` - `jaraco-classes ==3.2.2` - `jinja2 ==3.1.2` - `keyring ==23.9.1` - `markupsafe ==2.1.1` - `more-itertools ==8.14.0` - `packaging ==21.3` - `protobuf ==3.20.2` - `pyasn1 ==0.4.8` - `pyasn1-modules ==0.2.8` - `pycparser ==2.21` - `pyjwt ==2.4.0` - `pyparsing ==3.0.9` - `pyperclip ==1.8.2` - `python-dateutil ==2.8.2` - `pywin32-ctypes ==0.2.0` - `requests ==2.28.1` - `rsa ==4.9` - `six ==1.16.0` - `typing-extensions ==4.3.0` - `urllib3 ==1.26.12` - `zipp ==3.8.1` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull pull pull pull pull pull pull click on this checkbox to rebase all open prs at once ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull pull pull detected dependencies github actions github workflows build and test on windows yml actions checkout actions setup dotnet github workflows build and test yml actions checkout actions setup dotnet github workflows integration tests on emulator yml actions checkout actions setup dotnet github workflows integration tests on production yml actions checkout actions setup dotnet nuget google cloud spanner connection tests google cloud spanner connection tests csproj coverlet collector xunit runner visualstudio xunit combinatorial xunit microsoft net test sdk google cloud spanner connection google cloud spanner connection csproj google cloud spanner data google cloud spanner nhibernate benchmarks google cloud spanner nhibernate benchmarks csproj benchmarkdotnet google cloud spanner nhibernate integrationtests google cloud spanner nhibernate integrationtests csproj xunit skippablefact coverlet collector xunit runner visualstudio xunit combinatorial xunit microsoft net test sdk google cloud spanner nhibernate samples google cloud spanner nhibernate samples csproj docker dotnet google cloud spanner nhibernate tests google cloud spanner nhibernate tests csproj coverlet collector xunit runner visualstudio xunit combinatorial xunit microsoft net test sdk google cloud spanner nhibernate google cloud spanner nhibernate csproj nhibernate pip requirements kokoro requirements txt attrs cachetools certifi cffi charset normalizer click colorama cryptography gcp releasetool google auth idna importlib metadata jaraco classes keyring markupsafe more itertools packaging protobuf modules pycparser pyjwt pyparsing pyperclip python dateutil ctypes requests rsa six typing extensions zipp check this box to trigger a request for renovate to run again on this repository
1
691,291
23,691,105,218
IssuesEvent
2022-08-29 10:52:49
googleapis/python-pubsub
https://api.github.com/repos/googleapis/python-pubsub
closed
tests.system.TestStreamingPull: test_streaming_pull_ack_deadline failed
api: pubsub type: bug priority: p2 flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: b48a5a5dc43c95ce8c466686e568e72c583de2f4 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cc1da8e3-578e-4bf5-845a-55cc5db2e606), [Sponge](http://sponge2/cc1da8e3-578e-4bf5-845a-55cc5db2e606) status: failed <details><summary>Test output</summary><br><pre>self = <tests.system.TestStreamingPull object at 0x7f5a18fe8eb0> publisher = <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010> subscriber = <google.cloud.pubsub_v1.SubscriberClient object at 0x7f5a17738220> project = 'precise-truck-742' topic_path = 'projects/precise-truck-742/topics/t-1658003098583' subscription_path = 'projects/precise-truck-742/subscriptions/s-1658003098583' cleanup = [(<bound method PublisherClient.delete_topic of <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010>>, ()...erClient object at 0x7f5a17738220>>, (), {'subscription': 'projects/precise-truck-742/subscriptions/s-1658003098583'})] def test_streaming_pull_ack_deadline( self, publisher, subscriber, project, topic_path, subscription_path, cleanup ): # Make sure the topic and subscription get deleted. cleanup.append((publisher.delete_topic, (), {"topic": topic_path})) cleanup.append( (subscriber.delete_subscription, (), {"subscription": subscription_path}) ) # Create a topic and a subscription, then subscribe to the topic. This # must happen before the messages are published. publisher.create_topic(name=topic_path) # Subscribe to the topic. This must happen before the messages # are published. subscriber.create_subscription( name=subscription_path, topic=topic_path, ack_deadline_seconds=45 ) # publish some messages and wait for completion _publish_messages(publisher, topic_path, batch_sizes=[2]) # subscribe to the topic callback = StreamingPullCallback( processing_time=13, # more than the default stream ACK deadline (10s) resolve_at_msg_count=3, # one more than the published messages count ) flow_control = types.FlowControl(max_messages=1) subscription_future = subscriber.subscribe( subscription_path, callback, flow_control=flow_control ) # We expect to process the first two messages in 2 * 13 seconds, and # any duplicate message that is re-sent by the backend in additional # 13 seconds, totalling 39 seconds (+ overhead) --> if there have been # no duplicates in 60 seconds, we can reasonably assume that there # won't be any. try: callback.done_future.result(timeout=60) except exceptions.TimeoutError: # future timed out, because we received no excessive messages assert sorted(callback.seen_message_ids) == [1, 2] else: > pytest.fail( "Expected to receive 2 messages, but got at least {}.".format( len(callback.seen_message_ids) ) ) E Failed: Expected to receive 2 messages, but got at least 3. tests/system.py:552: Failed</pre></details>
1.0
tests.system.TestStreamingPull: test_streaming_pull_ack_deadline failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: b48a5a5dc43c95ce8c466686e568e72c583de2f4 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cc1da8e3-578e-4bf5-845a-55cc5db2e606), [Sponge](http://sponge2/cc1da8e3-578e-4bf5-845a-55cc5db2e606) status: failed <details><summary>Test output</summary><br><pre>self = <tests.system.TestStreamingPull object at 0x7f5a18fe8eb0> publisher = <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010> subscriber = <google.cloud.pubsub_v1.SubscriberClient object at 0x7f5a17738220> project = 'precise-truck-742' topic_path = 'projects/precise-truck-742/topics/t-1658003098583' subscription_path = 'projects/precise-truck-742/subscriptions/s-1658003098583' cleanup = [(<bound method PublisherClient.delete_topic of <google.cloud.pubsub_v1.PublisherClient object at 0x7f5a1773b010>>, ()...erClient object at 0x7f5a17738220>>, (), {'subscription': 'projects/precise-truck-742/subscriptions/s-1658003098583'})] def test_streaming_pull_ack_deadline( self, publisher, subscriber, project, topic_path, subscription_path, cleanup ): # Make sure the topic and subscription get deleted. cleanup.append((publisher.delete_topic, (), {"topic": topic_path})) cleanup.append( (subscriber.delete_subscription, (), {"subscription": subscription_path}) ) # Create a topic and a subscription, then subscribe to the topic. This # must happen before the messages are published. publisher.create_topic(name=topic_path) # Subscribe to the topic. This must happen before the messages # are published. subscriber.create_subscription( name=subscription_path, topic=topic_path, ack_deadline_seconds=45 ) # publish some messages and wait for completion _publish_messages(publisher, topic_path, batch_sizes=[2]) # subscribe to the topic callback = StreamingPullCallback( processing_time=13, # more than the default stream ACK deadline (10s) resolve_at_msg_count=3, # one more than the published messages count ) flow_control = types.FlowControl(max_messages=1) subscription_future = subscriber.subscribe( subscription_path, callback, flow_control=flow_control ) # We expect to process the first two messages in 2 * 13 seconds, and # any duplicate message that is re-sent by the backend in additional # 13 seconds, totalling 39 seconds (+ overhead) --> if there have been # no duplicates in 60 seconds, we can reasonably assume that there # won't be any. try: callback.done_future.result(timeout=60) except exceptions.TimeoutError: # future timed out, because we received no excessive messages assert sorted(callback.seen_message_ids) == [1, 2] else: > pytest.fail( "Expected to receive 2 messages, but got at least {}.".format( len(callback.seen_message_ids) ) ) E Failed: Expected to receive 2 messages, but got at least 3. tests/system.py:552: Failed</pre></details>
non_process
tests system teststreamingpull test streaming pull ack deadline failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output self publisher subscriber project precise truck topic path projects precise truck topics t subscription path projects precise truck subscriptions s cleanup def test streaming pull ack deadline self publisher subscriber project topic path subscription path cleanup make sure the topic and subscription get deleted cleanup append publisher delete topic topic topic path cleanup append subscriber delete subscription subscription subscription path create a topic and a subscription then subscribe to the topic this must happen before the messages are published publisher create topic name topic path subscribe to the topic this must happen before the messages are published subscriber create subscription name subscription path topic topic path ack deadline seconds publish some messages and wait for completion publish messages publisher topic path batch sizes subscribe to the topic callback streamingpullcallback processing time more than the default stream ack deadline resolve at msg count one more than the published messages count flow control types flowcontrol max messages subscription future subscriber subscribe subscription path callback flow control flow control we expect to process the first two messages in seconds and any duplicate message that is re sent by the backend in additional seconds totalling seconds overhead if there have been no duplicates in seconds we can reasonably assume that there won t be any try callback done future result timeout except exceptions timeouterror future timed out because we received no excessive messages assert sorted callback seen message ids else pytest fail expected to receive messages but got at least format len callback seen message ids e failed expected to receive messages but got at least tests system py failed
0
6,901
10,053,667,243
IssuesEvent
2019-07-21 18:43:56
tokio-rs/tokio
https://api.github.com/repos/tokio-rs/tokio
opened
process: Investigate using job objects and IOCP on Windows
tokio-process
Originally reported in alexcrichton/tokio-process#11: > While reading the MSDN docs on job objects for something, I came across this page: https://msdn.microsoft.com/en-us/library/windows/desktop/ms684141(v=vs.85).aspx > > If you create a job object, you can associate an IOCP with it. If you then create processes in that job (generally using [`AssignProcessToJobObject`](https://msdn.microsoft.com/en-us/library/windows/desktop/ms681949(v=vs.85).aspx) on a process that has been created with the `CREATE_SUSPENDED` flag) you will get process event messages on the IOCP, including `JOB_OBJECT_MSG_EXIT_PROCESS`. Please see the comments in the original discussion for context. It appears that this depends on tokio-rs/mio#526
1.0
process: Investigate using job objects and IOCP on Windows - Originally reported in alexcrichton/tokio-process#11: > While reading the MSDN docs on job objects for something, I came across this page: https://msdn.microsoft.com/en-us/library/windows/desktop/ms684141(v=vs.85).aspx > > If you create a job object, you can associate an IOCP with it. If you then create processes in that job (generally using [`AssignProcessToJobObject`](https://msdn.microsoft.com/en-us/library/windows/desktop/ms681949(v=vs.85).aspx) on a process that has been created with the `CREATE_SUSPENDED` flag) you will get process event messages on the IOCP, including `JOB_OBJECT_MSG_EXIT_PROCESS`. Please see the comments in the original discussion for context. It appears that this depends on tokio-rs/mio#526
process
process investigate using job objects and iocp on windows originally reported in alexcrichton tokio process while reading the msdn docs on job objects for something i came across this page if you create a job object you can associate an iocp with it if you then create processes in that job generally using on a process that has been created with the create suspended flag you will get process event messages on the iocp including job object msg exit process please see the comments in the original discussion for context it appears that this depends on tokio rs mio
1
654,767
21,662,157,185
IssuesEvent
2022-05-06 20:36:28
bounswe/bounswe2022group7
https://api.github.com/repos/bounswe/bounswe2022group7
closed
[Database] Implement "ArtItem", "DiscussionPost", and "Comment" models
Status: Completed Priority: Medium Difficulty: Medium Type: Implementation
[\_\_init.py\_\_](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/__init__.py) and [models.py](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/models.py) files are created with fundamental functions. Given those, I shall implement `ArtItem`, `DiscussionPost`, and `Comment` models. - Implement mentioned models - Push the updates to the [practice-app/database-base](https://github.com/bounswe/bounswe2022group7/tree/practice-app/database-base) branch **Deadline: 06/05/2022, 14:00** **Reviewer: @azizamankenova**
1.0
[Database] Implement "ArtItem", "DiscussionPost", and "Comment" models - [\_\_init.py\_\_](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/__init__.py) and [models.py](https://github.com/bounswe/bounswe2022group7/blob/practice-app/database-base/practice-app/website/models.py) files are created with fundamental functions. Given those, I shall implement `ArtItem`, `DiscussionPost`, and `Comment` models. - Implement mentioned models - Push the updates to the [practice-app/database-base](https://github.com/bounswe/bounswe2022group7/tree/practice-app/database-base) branch **Deadline: 06/05/2022, 14:00** **Reviewer: @azizamankenova**
non_process
implement artitem discussionpost and comment models and files are created with fundamental functions given those i shall implement artitem discussionpost and comment models implement mentioned models push the updates to the branch deadline reviewer azizamankenova
0
443,781
12,799,552,447
IssuesEvent
2020-07-02 15:34:20
bitnami-labs/kubewatch
https://api.github.com/repos/bitnami-labs/kubewatch
opened
Support watching arbitrary custom resources
enhancement priority-P1
Currently kubewatch supports only watching a built-in list of resource kinds. Nowadays end users consume many custom resources (CRs) and also new k8s versions keep adding new resources and we shouldn't need to update kubewatch every time such a new resource pops up.
1.0
Support watching arbitrary custom resources - Currently kubewatch supports only watching a built-in list of resource kinds. Nowadays end users consume many custom resources (CRs) and also new k8s versions keep adding new resources and we shouldn't need to update kubewatch every time such a new resource pops up.
non_process
support watching arbitrary custom resources currently kubewatch supports only watching a built in list of resource kinds nowadays end users consume many custom resources crs and also new versions keep adding new resources and we shouldn t need to update kubewatch every time such a new resource pops up
0
16,021
20,188,228,740
IssuesEvent
2022-02-11 01:19:51
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Develop a security training program
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Operational Model & DevOps Roles & Responsibilities
<a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/">Develop a security training program</a> <p><b>Why Consider This?</b></p> Cybersecurity threats are always evolving and therefore those responsible for organizational information security require specialized, continual, and relevant training to ensure staff maintains the level of competency required to protect, detect, and respond. <p><b>Context</b></p> <p><span>A security training program is a component of an overall comprehensive cybersecurity program."nbsp; End users are often the target of cyber attacks and therefore need to be equipped with the knowledge and tools to help protect the organization."nbsp; Security program development and maintenance requires on-going investment and support from leadership to make it a critical component of the organization."nbsp; </span></p> <p><b>Suggested Actions</b></p> <p><span>See </span><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span> for more information.</span></p> <p><b>Learn More</b></p> <p><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span /></p>
1.0
Develop a security training program - <a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/">Develop a security training program</a> <p><b>Why Consider This?</b></p> Cybersecurity threats are always evolving and therefore those responsible for organizational information security require specialized, continual, and relevant training to ensure staff maintains the level of competency required to protect, detect, and respond. <p><b>Context</b></p> <p><span>A security training program is a component of an overall comprehensive cybersecurity program."nbsp; End users are often the target of cyber attacks and therefore need to be equipped with the knowledge and tools to help protect the organization."nbsp; Security program development and maintenance requires on-going investment and support from leadership to make it a critical component of the organization."nbsp; </span></p> <p><b>Suggested Actions</b></p> <p><span>See </span><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span> for more information.</span></p> <p><b>Learn More</b></p> <p><a href="https://www.microsoft.com/itshowcase/blog/how-microsoft-is-transforming-its-approach-to-security-training/" target="_blank"><span>How Microsoft is transforming its approach to security training</span></a><span /></p>
process
develop a security training program why consider this cybersecurity threats are always evolving and therefore those responsible for organizational information security require specialized continual and relevant training to ensure staff maintains the level of competency required to protect detect and respond context a security training program is a component of an overall comprehensive cybersecurity program nbsp end users are often the target of cyber attacks and therefore need to be equipped with the knowledge and tools to help protect the organization nbsp security program development and maintenance requires on going investment and support from leadership to make it a critical component of the organization nbsp suggested actions see how microsoft is transforming its approach to security training for more information learn more how microsoft is transforming its approach to security training
1
143,276
19,177,907,851
IssuesEvent
2021-12-04 00:04:02
samq-ghdemo/js-monorepo
https://api.github.com/repos/samq-ghdemo/js-monorepo
opened
CVE-2014-10064 (High) detected in qs-0.6.6.tgz
security vulnerability
## CVE-2014-10064 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.6.6.tgz</b></p></summary> <p>querystring parser</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p> <p>Path to dependency file: js-monorepo/NodeGoat/package.json</p> <p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/zaproxy/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - request-2.36.0.tgz - :x: **qs-0.6.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064>CVE-2014-10064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-10064">https://nvd.nist.gov/vuln/detail/CVE-2014-10064</a></p> <p>Release Date: 2018-05-31</p> <p>Fix Resolution: qs - 1.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 1.0.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2014-10064","vulnerabilityDetails":"The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2014-10064 (High) detected in qs-0.6.6.tgz - ## CVE-2014-10064 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.6.6.tgz</b></p></summary> <p>querystring parser</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p> <p>Path to dependency file: js-monorepo/NodeGoat/package.json</p> <p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/zaproxy/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - request-2.36.0.tgz - :x: **qs-0.6.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064>CVE-2014-10064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-10064">https://nvd.nist.gov/vuln/detail/CVE-2014-10064</a></p> <p>Release Date: 2018-05-31</p> <p>Fix Resolution: qs - 1.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 1.0.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2014-10064","vulnerabilityDetails":"The qs module before 1.0.0 does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time. An attacker could leverage this to cause a temporary denial-of-service condition, for example, in a web application, other requests would not be processed while this blocking is occurring.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-10064","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz querystring parser library home page a href path to dependency file js monorepo nodegoat package json path to vulnerable library js monorepo nodegoat node modules zaproxy node modules qs package json dependency hierarchy zaproxy tgz root library request tgz x qs tgz vulnerable library found in head commit a href found in base branch main vulnerability details the qs module before does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time an attacker could leverage this to cause a temporary denial of service condition for example in a web application other requests would not be processed while this blocking is occurring publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree zaproxy request qs isminimumfixversionavailable true minimumfixversion qs isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the qs module before does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time an attacker could leverage this to cause a temporary denial of service condition for example in a web application other requests would not be processed while this blocking is occurring vulnerabilityurl
0
17,067
22,505,126,751
IssuesEvent
2022-06-23 14:54:14
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[API Proposal]: `Process.Architecture` or `Process.Is64Bit`
api-suggestion area-System.Diagnostics.Process
### Background and motivation Working a lot with external process memory, I find it a little tedious having to use [`IsWow64Process`](https://docs.microsoft.com/en-us/windows/win32/api/wow64apiset/nf-wow64apiset-iswow64process) each time I want to infer the architecture of a target process. Not to mention, that function's `out Wow64Process` parameter is heaps confusing. It would be a lot nicer if I could simply check a `Process.Is64Bit` property. ### API Proposal ```cs namespace System.Diagnostics; public class Process { public bool Is64Bit { get; } } ``` ### API Usage ```csharp var proc = Process.GetCurrentProcess(); Console.WriteLine(proc.Is64Bit); ``` ### Alternative Designs ```cs namespace System.Diagnostics; public class Process { public enum ProcessArchitecture { x64, x86, TheFutureIsNow } public ProcessArchitecture Architecture { get; } } ``` ### Risks _No response_
1.0
[API Proposal]: `Process.Architecture` or `Process.Is64Bit` - ### Background and motivation Working a lot with external process memory, I find it a little tedious having to use [`IsWow64Process`](https://docs.microsoft.com/en-us/windows/win32/api/wow64apiset/nf-wow64apiset-iswow64process) each time I want to infer the architecture of a target process. Not to mention, that function's `out Wow64Process` parameter is heaps confusing. It would be a lot nicer if I could simply check a `Process.Is64Bit` property. ### API Proposal ```cs namespace System.Diagnostics; public class Process { public bool Is64Bit { get; } } ``` ### API Usage ```csharp var proc = Process.GetCurrentProcess(); Console.WriteLine(proc.Is64Bit); ``` ### Alternative Designs ```cs namespace System.Diagnostics; public class Process { public enum ProcessArchitecture { x64, x86, TheFutureIsNow } public ProcessArchitecture Architecture { get; } } ``` ### Risks _No response_
process
process architecture or process background and motivation working a lot with external process memory i find it a little tedious having to use each time i want to infer the architecture of a target process not to mention that function s out parameter is heaps confusing it would be a lot nicer if i could simply check a process property api proposal cs namespace system diagnostics public class process public bool get api usage csharp var proc process getcurrentprocess console writeline proc alternative designs cs namespace system diagnostics public class process public enum processarchitecture thefutureisnow public processarchitecture architecture get risks no response
1
433,331
12,505,688,934
IssuesEvent
2020-06-02 11:12:10
gitcoinco/web
https://api.github.com/repos/gitcoinco/web
closed
Wonky layout in the "x new funded issues" mails
bug priority: backlog
**Describe the bug** Layout between the "funder" and the text is all over the place. **To Reproduce** Receive e-mail, read e-mail **Expected behavior** A clear line between "owner"/"tasker"/"requester" and text **Screenshots** ![image](https://user-images.githubusercontent.com/761911/47884535-71465a80-de31-11e8-995a-5176092151f7.png) **Desktop (please complete the following information):** - OS: Mac OS X - Mail Client: Spark
1.0
Wonky layout in the "x new funded issues" mails - **Describe the bug** Layout between the "funder" and the text is all over the place. **To Reproduce** Receive e-mail, read e-mail **Expected behavior** A clear line between "owner"/"tasker"/"requester" and text **Screenshots** ![image](https://user-images.githubusercontent.com/761911/47884535-71465a80-de31-11e8-995a-5176092151f7.png) **Desktop (please complete the following information):** - OS: Mac OS X - Mail Client: Spark
non_process
wonky layout in the x new funded issues mails describe the bug layout between the funder and the text is all over the place to reproduce receive e mail read e mail expected behavior a clear line between owner tasker requester and text screenshots desktop please complete the following information os mac os x mail client spark
0
21,505
29,670,497,061
IssuesEvent
2023-06-11 10:49:20
giorgiopizz/mkShapesRDF
https://api.github.com/repos/giorgiopizz/mkShapesRDF
closed
[PROC] [SHAPE] Prepare Latinos Hands-On tutorial
shape-analysis processor
The tutorial should be composed of two parts, both working with UL 2018v9: 1. Using processor, define two reasonable workflows for a double muon dataset and DY NLO 2. Using shape analysis create another example folder that works with latinos root files (as it is for 2016Real) that makes shapes for the DY CR in the double muon channel
1.0
[PROC] [SHAPE] Prepare Latinos Hands-On tutorial - The tutorial should be composed of two parts, both working with UL 2018v9: 1. Using processor, define two reasonable workflows for a double muon dataset and DY NLO 2. Using shape analysis create another example folder that works with latinos root files (as it is for 2016Real) that makes shapes for the DY CR in the double muon channel
process
prepare latinos hands on tutorial the tutorial should be composed of two parts both working with ul using processor define two reasonable workflows for a double muon dataset and dy nlo using shape analysis create another example folder that works with latinos root files as it is for that makes shapes for the dy cr in the double muon channel
1
108,761
9,331,088,699
IssuesEvent
2019-03-28 08:56:31
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits Test failure
:Search/Search >test-failure v7.1.0 v8.0.0
Fails reliably on master and 7.x (possibly other branches) ``` ./gradlew :server:unitTest -Dtests.seed=6AD7B08D850920A6 -Dtests.class=org.elasticsearch.action.search.SearchResponseMergerTests -Dtests.method="testMergeSearchHits" -Dtests.security.manager=true -Dtests.locale=ar-SY -Dtests.timezone=SystemV/MST7 -Dcompiler.java=12 -Druntime.java=8 ``` Failure: ``` FAILURE 0.67s | SearchResponseMergerTests.testMergeSearchHits <<< FAILURES! > Throwable #1: java.lang.AssertionError: expected null, but was:<[Lorg.apache.lucene.search.SortField;@6c4ebb78> > at __randomizedtesting.SeedInfo.seed([6AD7B08D850920A6:3AC351382A6A6789]:0) > at org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits(SearchResponseMergerTests.java:520) > at java.lang.Thread.run(Thread.java:748) ```
1.0
org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits Test failure - Fails reliably on master and 7.x (possibly other branches) ``` ./gradlew :server:unitTest -Dtests.seed=6AD7B08D850920A6 -Dtests.class=org.elasticsearch.action.search.SearchResponseMergerTests -Dtests.method="testMergeSearchHits" -Dtests.security.manager=true -Dtests.locale=ar-SY -Dtests.timezone=SystemV/MST7 -Dcompiler.java=12 -Druntime.java=8 ``` Failure: ``` FAILURE 0.67s | SearchResponseMergerTests.testMergeSearchHits <<< FAILURES! > Throwable #1: java.lang.AssertionError: expected null, but was:<[Lorg.apache.lucene.search.SortField;@6c4ebb78> > at __randomizedtesting.SeedInfo.seed([6AD7B08D850920A6:3AC351382A6A6789]:0) > at org.elasticsearch.action.search.SearchResponseMergerTests.testMergeSearchHits(SearchResponseMergerTests.java:520) > at java.lang.Thread.run(Thread.java:748) ```
non_process
org elasticsearch action search searchresponsemergertests testmergesearchhits test failure fails reliably on master and x possibly other branches gradlew server unittest dtests seed dtests class org elasticsearch action search searchresponsemergertests dtests method testmergesearchhits dtests security manager true dtests locale ar sy dtests timezone systemv dcompiler java druntime java failure failure searchresponsemergertests testmergesearchhits failures throwable java lang assertionerror expected null but was at randomizedtesting seedinfo seed at org elasticsearch action search searchresponsemergertests testmergesearchhits searchresponsemergertests java at java lang thread run thread java
0
12,609
15,012,974,343
IssuesEvent
2021-02-01 03:07:08
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
closed
Recommended checkbox functionality
FE ShapeupProcess challenge- recommender-tool
Recommended checkbox must not be part of the challenge type group. It must a standalone checkbox <img width="436" alt="Screenshot 2021-01-27 at 3 37 03 PM" src="https://user-images.githubusercontent.com/58783823/105976274-f757b700-60b5-11eb-89ae-5fcd8b17a4e6.png"> At present checking the recommended checkbox, deselects the challenge type checkboxes. This must not be the case.
1.0
Recommended checkbox functionality - Recommended checkbox must not be part of the challenge type group. It must a standalone checkbox <img width="436" alt="Screenshot 2021-01-27 at 3 37 03 PM" src="https://user-images.githubusercontent.com/58783823/105976274-f757b700-60b5-11eb-89ae-5fcd8b17a4e6.png"> At present checking the recommended checkbox, deselects the challenge type checkboxes. This must not be the case.
process
recommended checkbox functionality recommended checkbox must not be part of the challenge type group it must a standalone checkbox img width alt screenshot at pm src at present checking the recommended checkbox deselects the challenge type checkboxes this must not be the case
1
480,008
13,821,856,508
IssuesEvent
2020-10-13 03:31:55
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
No voting for removing someone from office
Category: Laws Priority: Medium Type: Feature
It seems like the supreme court that can remove a president from office can not do so by a majority vote of the three, but the members are only able to remove him from office by themselves - each one can alone.
1.0
No voting for removing someone from office - It seems like the supreme court that can remove a president from office can not do so by a majority vote of the three, but the members are only able to remove him from office by themselves - each one can alone.
non_process
no voting for removing someone from office it seems like the supreme court that can remove a president from office can not do so by a majority vote of the three but the members are only able to remove him from office by themselves each one can alone
0
5,476
27,363,850,307
IssuesEvent
2023-02-27 17:37:51
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Bug: samconfig.toml values don't override default values in sam template
blocked/more-info-needed blocked/close-if-inactive maintainer/need-followup
### Description: Running a .NET Core 3.1 project containing Lambdas, trying to deploy using sam build and sam deploy. Build is successful, but Deploy fails in AWS because parameter overrides specified in samconfig.toml are not used, even though log for both sam build and sam deploy states that they are found correctly. ### Steps to reproduce: Running ADO pipeline (attached), deploying SAM template (attached) with samconfig.toml. ### Observed result: The template generated by sam build has original default values specified for parameters, and they are the ones used during deployment, even though log says : for sam build: Loading configuration values from [uat.['build'].parameters] (env.command_name.section) in config file at '/workspace/1/s/Main/Source/Services/samconfig.toml' and similar for sam deploy ### Expected result: the template to be deployed with parameters specified in [uat.global.parameters] section in samconfig.toml ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) OS: Lambda in AWS sam --version: Tool 'amazon.lambda.tools' (version '5.6.2') AWS region: us-west-2 Attached are logs for both Build and Deploy steps as well as the original template, generated one and samconfig.toml (in templatesAndSamCfg.zip) [Build-Log.txt](https://github.com/aws/aws-sam-cli/files/10406008/Build-Log.txt) [Deploy-Log.txt](https://github.com/aws/aws-sam-cli/files/10406009/Deploy-Log.txt) [samconfig.zip](https://github.com/aws/aws-sam-cli/files/10406011/samconfig.zip) [templatesAndSamCfg.zip](https://github.com/aws/aws-sam-cli/files/10406018/templatesAndSamCfg.zip)
True
Bug: samconfig.toml values don't override default values in sam template - ### Description: Running a .NET Core 3.1 project containing Lambdas, trying to deploy using sam build and sam deploy. Build is successful, but Deploy fails in AWS because parameter overrides specified in samconfig.toml are not used, even though log for both sam build and sam deploy states that they are found correctly. ### Steps to reproduce: Running ADO pipeline (attached), deploying SAM template (attached) with samconfig.toml. ### Observed result: The template generated by sam build has original default values specified for parameters, and they are the ones used during deployment, even though log says : for sam build: Loading configuration values from [uat.['build'].parameters] (env.command_name.section) in config file at '/workspace/1/s/Main/Source/Services/samconfig.toml' and similar for sam deploy ### Expected result: the template to be deployed with parameters specified in [uat.global.parameters] section in samconfig.toml ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) OS: Lambda in AWS sam --version: Tool 'amazon.lambda.tools' (version '5.6.2') AWS region: us-west-2 Attached are logs for both Build and Deploy steps as well as the original template, generated one and samconfig.toml (in templatesAndSamCfg.zip) [Build-Log.txt](https://github.com/aws/aws-sam-cli/files/10406008/Build-Log.txt) [Deploy-Log.txt](https://github.com/aws/aws-sam-cli/files/10406009/Deploy-Log.txt) [samconfig.zip](https://github.com/aws/aws-sam-cli/files/10406011/samconfig.zip) [templatesAndSamCfg.zip](https://github.com/aws/aws-sam-cli/files/10406018/templatesAndSamCfg.zip)
non_process
bug samconfig toml values don t override default values in sam template description running a net core project containing lambdas trying to deploy using sam build and sam deploy build is successful but deploy fails in aws because parameter overrides specified in samconfig toml are not used even though log for both sam build and sam deploy states that they are found correctly steps to reproduce running ado pipeline attached deploying sam template attached with samconfig toml observed result the template generated by sam build has original default values specified for parameters and they are the ones used during deployment even though log says for sam build loading configuration values from parameters env command name section in config file at workspace s main source services samconfig toml and similar for sam deploy expected result the template to be deployed with parameters specified in section in samconfig toml additional environment details ex windows mac amazon linux etc os lambda in aws sam version tool amazon lambda tools version aws region us west attached are logs for both build and deploy steps as well as the original template generated one and samconfig toml in templatesandsamcfg zip
0
275,299
23,904,726,369
IssuesEvent
2022-09-08 22:44:55
handsontable/hyperformula
https://api.github.com/repos/handsontable/hyperformula
closed
Run performance benchmark automatically for each PR
Chore Tests Released Performance Impact: Medium
### Description <!--- [mandatory] Describe the actual behavior and expected behavior --> Run performance benchmark automatically for each PR. Something similar is done in Spreadsheet Viewer repo. It might be helpful ### Links - https://github.com/handsontable/hyperformula/pull/938#issuecomment-1090052862 - https://handsoncode.slack.com/archives/D02T462B3GS/p1648823588399299 (internal discussion - accessible by @sequba )
1.0
Run performance benchmark automatically for each PR - ### Description <!--- [mandatory] Describe the actual behavior and expected behavior --> Run performance benchmark automatically for each PR. Something similar is done in Spreadsheet Viewer repo. It might be helpful ### Links - https://github.com/handsontable/hyperformula/pull/938#issuecomment-1090052862 - https://handsoncode.slack.com/archives/D02T462B3GS/p1648823588399299 (internal discussion - accessible by @sequba )
non_process
run performance benchmark automatically for each pr description run performance benchmark automatically for each pr something similar is done in spreadsheet viewer repo it might be helpful links internal discussion accessible by sequba
0
169,901
13,164,828,075
IssuesEvent
2020-08-11 04:57:04
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
FNX-12780 ⁃ Intermittent UI test failure - defaultDesktopBookmarksFoldersTest
eng:ui-test intermittent-test
### Firebase Test Run: https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7976327018966179987/executions/bs.58e6f78a50568c44/testcases/1/test-cases https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7618900378390394025/executions/bs.2f5d0fab0c8245d1/testcases/1/test-cases ### Stacktrace: `androidx.test.espresso.base.DefaultFailureHandler$AssertionFailedWithCauseError: 'is displayed on the screen to the user' doesn't match the selected view. Expected: is displayed on the screen to the user Got: "MaterialButton{id=2131362047, res-name=bookmark_folders_sign_in, visibility=VISIBLE, width=992, height=132, has-focus=false, has-focusable=true, has-window-focus=true, is-clickable=true, is-enabled=true, is-focused=false, is-focusable=true, is-layout-requested=false, is-selected=false, layout-params=android.widget.LinearLayout$LayoutParams@dc6d387, tag=null, root-is-layout-requested=false, has-input-connection=false, x=44.0, y=1568.0, text=Sign in to see synced bookmarks, input-type=0, ime-target=false, has-links=false, is-checked=false}" at dalvik.system.VMStack.getThreadStackTrace(Native Method) at java.lang.Thread.getStackTrace(Thread.java:1538) at androidx.test.espresso.base.DefaultFailureHandler.getUserFriendlyError(DefaultFailureHandler.java:16) at androidx.test.espresso.base.DefaultFailureHandler.handle(DefaultFailureHandler.java:36) at androidx.test.espresso.ViewInteraction.waitForAndHandleInteractionResults(ViewInteraction.java:103) at androidx.test.espresso.ViewInteraction.check(ViewInteraction.java:31) at org.mozilla.fenix.ui.robots.BookmarksRobot.verifySignInToSyncButton(BookmarksRobot.kt:116) at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:85) at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:35) at org.mozilla.fenix.ui.robots.ThreeDotMenuMainRobot$Transition.openBookmarks(ThreeDotMenuMainRobot.kt:128) at org.mozilla.fenix.ui.BookmarksTest.defaultDesktopBookmarksFoldersTest(BookmarksTest.kt:76)` ### Build: 6/18/20, 6/19/20
2.0
FNX-12780 ⁃ Intermittent UI test failure - defaultDesktopBookmarksFoldersTest - ### Firebase Test Run: https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7976327018966179987/executions/bs.58e6f78a50568c44/testcases/1/test-cases https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7618900378390394025/executions/bs.2f5d0fab0c8245d1/testcases/1/test-cases ### Stacktrace: `androidx.test.espresso.base.DefaultFailureHandler$AssertionFailedWithCauseError: 'is displayed on the screen to the user' doesn't match the selected view. Expected: is displayed on the screen to the user Got: "MaterialButton{id=2131362047, res-name=bookmark_folders_sign_in, visibility=VISIBLE, width=992, height=132, has-focus=false, has-focusable=true, has-window-focus=true, is-clickable=true, is-enabled=true, is-focused=false, is-focusable=true, is-layout-requested=false, is-selected=false, layout-params=android.widget.LinearLayout$LayoutParams@dc6d387, tag=null, root-is-layout-requested=false, has-input-connection=false, x=44.0, y=1568.0, text=Sign in to see synced bookmarks, input-type=0, ime-target=false, has-links=false, is-checked=false}" at dalvik.system.VMStack.getThreadStackTrace(Native Method) at java.lang.Thread.getStackTrace(Thread.java:1538) at androidx.test.espresso.base.DefaultFailureHandler.getUserFriendlyError(DefaultFailureHandler.java:16) at androidx.test.espresso.base.DefaultFailureHandler.handle(DefaultFailureHandler.java:36) at androidx.test.espresso.ViewInteraction.waitForAndHandleInteractionResults(ViewInteraction.java:103) at androidx.test.espresso.ViewInteraction.check(ViewInteraction.java:31) at org.mozilla.fenix.ui.robots.BookmarksRobot.verifySignInToSyncButton(BookmarksRobot.kt:116) at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:85) at org.mozilla.fenix.ui.BookmarksTest$defaultDesktopBookmarksFoldersTest$3.invoke(BookmarksTest.kt:35) at org.mozilla.fenix.ui.robots.ThreeDotMenuMainRobot$Transition.openBookmarks(ThreeDotMenuMainRobot.kt:128) at org.mozilla.fenix.ui.BookmarksTest.defaultDesktopBookmarksFoldersTest(BookmarksTest.kt:76)` ### Build: 6/18/20, 6/19/20
non_process
fnx ⁃ intermittent ui test failure defaultdesktopbookmarksfolderstest firebase test run stacktrace androidx test espresso base defaultfailurehandler assertionfailedwithcauseerror is displayed on the screen to the user doesn t match the selected view expected is displayed on the screen to the user got materialbutton id res name bookmark folders sign in visibility visible width height has focus false has focusable true has window focus true is clickable true is enabled true is focused false is focusable true is layout requested false is selected false layout params android widget linearlayout layoutparams tag null root is layout requested false has input connection false x y text sign in to see synced bookmarks input type ime target false has links false is checked false at dalvik system vmstack getthreadstacktrace native method at java lang thread getstacktrace thread java at androidx test espresso base defaultfailurehandler getuserfriendlyerror defaultfailurehandler java at androidx test espresso base defaultfailurehandler handle defaultfailurehandler java at androidx test espresso viewinteraction waitforandhandleinteractionresults viewinteraction java at androidx test espresso viewinteraction check viewinteraction java at org mozilla fenix ui robots bookmarksrobot verifysignintosyncbutton bookmarksrobot kt at org mozilla fenix ui bookmarkstest defaultdesktopbookmarksfolderstest invoke bookmarkstest kt at org mozilla fenix ui bookmarkstest defaultdesktopbookmarksfolderstest invoke bookmarkstest kt at org mozilla fenix ui robots threedotmenumainrobot transition openbookmarks threedotmenumainrobot kt at org mozilla fenix ui bookmarkstest defaultdesktopbookmarksfolderstest bookmarkstest kt build
0
12,480
14,949,074,403
IssuesEvent
2021-01-26 10:59:17
cmux/koot
https://api.github.com/repos/cmux/koot
closed
新的生命周期 beforeBuild & afterBuild
bundling process enhancement
```koot.config.js module.exports = { // ... beforeBuild: async (fullConfig) => {}, afterBuild: async (fullConfig) => {}, // ... }; ```
1.0
新的生命周期 beforeBuild & afterBuild - ```koot.config.js module.exports = { // ... beforeBuild: async (fullConfig) => {}, afterBuild: async (fullConfig) => {}, // ... }; ```
process
新的生命周期 beforebuild afterbuild koot config js module exports beforebuild async fullconfig afterbuild async fullconfig
1
14,351
17,374,572,807
IssuesEvent
2021-07-30 18:49:41
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Unlink functionnality actually seem to apply to release pipelines
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
> The Link and Unlink functionality applies to build pipelines only. It does not apply to release pipelines. The "Unlink all" button does exist on "Releases" Pipelines, and is referenced on [another doc](https://docs.microsoft.com/en-us/azure/devops/pipelines/library/task-groups?view=azure-devops). I personally don't see it on a "Pipelines" Pipeline, or what we would call "Build" Pipelines. But then, what is refered as Build Pipelines in the docs are actually under "Pipelines" > "Pipelines" in the Azure DevOps menu, so I can't state with certainty what you are referring to when you say "build pipelines" on this page (I find this to be very confusing and I won't miss an occasion to mention it ^^). --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 1ec7e5a9-e22e-a329-14cf-6e21e65cc85f * Version Independent ID: c96aaf4e-f5ec-c42f-05cb-e2366167fbc6 * Content: [Process parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/parameters?view=azure-devops) * Content Source: [docs/pipelines/process/parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/parameters.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Unlink functionnality actually seem to apply to release pipelines - > The Link and Unlink functionality applies to build pipelines only. It does not apply to release pipelines. The "Unlink all" button does exist on "Releases" Pipelines, and is referenced on [another doc](https://docs.microsoft.com/en-us/azure/devops/pipelines/library/task-groups?view=azure-devops). I personally don't see it on a "Pipelines" Pipeline, or what we would call "Build" Pipelines. But then, what is refered as Build Pipelines in the docs are actually under "Pipelines" > "Pipelines" in the Azure DevOps menu, so I can't state with certainty what you are referring to when you say "build pipelines" on this page (I find this to be very confusing and I won't miss an occasion to mention it ^^). --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 1ec7e5a9-e22e-a329-14cf-6e21e65cc85f * Version Independent ID: c96aaf4e-f5ec-c42f-05cb-e2366167fbc6 * Content: [Process parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/parameters?view=azure-devops) * Content Source: [docs/pipelines/process/parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/parameters.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
unlink functionnality actually seem to apply to release pipelines the link and unlink functionality applies to build pipelines only it does not apply to release pipelines the unlink all button does exist on releases pipelines and is referenced on i personally don t see it on a pipelines pipeline or what we would call build pipelines but then what is refered as build pipelines in the docs are actually under pipelines pipelines in the azure devops menu so i can t state with certainty what you are referring to when you say build pipelines on this page i find this to be very confusing and i won t miss an occasion to mention it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
9,667
12,675,347,071
IssuesEvent
2020-06-19 01:23:21
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Still don't know where Run (build) number is set in classic pipelines
Pri1 devops-cicd-process/tech devops/prod
I came to this page to find out where the "Build Number" option moved since I can no longer find it to edit in our classic pipelines. After reading this page, I still have no idea where the actual field is (despite a note saying that you can leave it blank). If it's gone, it should be documented here. If it's supposed to be in the new UI but it's not, there's a bug in the UI and an issue should be opened there. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Still don't know where Run (build) number is set in classic pipelines - I came to this page to find out where the "Build Number" option moved since I can no longer find it to edit in our classic pipelines. After reading this page, I still have no idea where the actual field is (despite a note saying that you can leave it blank). If it's gone, it should be documented here. If it's supposed to be in the new UI but it's not, there's a bug in the UI and an issue should be opened there. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
still don t know where run build number is set in classic pipelines i came to this page to find out where the build number option moved since i can no longer find it to edit in our classic pipelines after reading this page i still have no idea where the actual field is despite a note saying that you can leave it blank if it s gone it should be documented here if it s supposed to be in the new ui but it s not there s a bug in the ui and an issue should be opened there document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
77,781
7,603,563,912
IssuesEvent
2018-04-29 15:48:58
OregonStateRocketry/30k2018-CS-Capstone
https://api.github.com/repos/OregonStateRocketry/30k2018-CS-Capstone
closed
Avionics - rocket is using payload's state machine
before practice launch #1 required tests
Right now mainRocket is using the state machine designed for mainPayload. This will not work because the main rocket does not go through the same phases. Many stages are the same but others are not. It might be confusing to use logic to separate them. I think we should probably make sure it does what we want it to, then clone this class and customize it for the rocket phases.
1.0
Avionics - rocket is using payload's state machine - Right now mainRocket is using the state machine designed for mainPayload. This will not work because the main rocket does not go through the same phases. Many stages are the same but others are not. It might be confusing to use logic to separate them. I think we should probably make sure it does what we want it to, then clone this class and customize it for the rocket phases.
non_process
avionics rocket is using payload s state machine right now mainrocket is using the state machine designed for mainpayload this will not work because the main rocket does not go through the same phases many stages are the same but others are not it might be confusing to use logic to separate them i think we should probably make sure it does what we want it to then clone this class and customize it for the rocket phases
0
593,945
18,020,551,057
IssuesEvent
2021-09-16 18:50:16
google/shaka-player
https://api.github.com/repos/google/shaka-player
closed
Networking engine does not provide response HTTP status code
enhancement contributions welcome Why didn't we catch this sooner? priority:P3
**Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?** Yes **What version of Shaka Player are you using?** 3.2.0 **Can you reproduce the issue with our latest release version?** Yes **Can you reproduce the issue with the latest code from `master`?** Yes **Are you using the demo app or your own custom app?** Irrelevant **If custom app, can you reproduce the issue using our demo app?** Irrelevant **What browser and OS are you using?** Irrelevant **For embedded devices (smart TVs, etc.), what model and firmware version are you using?** Irrelevant **What are the manifest and license server URIs?** Irrelevant **What configuration are you using? What is the output of `player.getConfiguration()`?** Irrelevant **What did you do?** I was trying to implement "**License Wrapping**", based on the official tutorial, as the result I had the following code: ```js player.getNetworkingEngine().registerResponseFilter(function (type, response) { if (!type == shaka.net.NetworkingEngine.RequestType.LICENSE) return; if (response.status === 200) ...One scenario if (response.status === 201) ...Another scenario }); ``` Unfortunately, my code did not work and it took some time in order to find out the reason. It seems that _Networking Engine_ does not provide **Response Status** which is weird, because it provides almost all the information. **What did you expect to happen?** I was expecting that **response** will include **response.status**. **What actually happened?** As I mentioned before, my License Wrapping implementation did not work, because I was not able to compare response.status. I don't know the reason why it was excluded from response object.
1.0
Networking engine does not provide response HTTP status code - **Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?** Yes **What version of Shaka Player are you using?** 3.2.0 **Can you reproduce the issue with our latest release version?** Yes **Can you reproduce the issue with the latest code from `master`?** Yes **Are you using the demo app or your own custom app?** Irrelevant **If custom app, can you reproduce the issue using our demo app?** Irrelevant **What browser and OS are you using?** Irrelevant **For embedded devices (smart TVs, etc.), what model and firmware version are you using?** Irrelevant **What are the manifest and license server URIs?** Irrelevant **What configuration are you using? What is the output of `player.getConfiguration()`?** Irrelevant **What did you do?** I was trying to implement "**License Wrapping**", based on the official tutorial, as the result I had the following code: ```js player.getNetworkingEngine().registerResponseFilter(function (type, response) { if (!type == shaka.net.NetworkingEngine.RequestType.LICENSE) return; if (response.status === 200) ...One scenario if (response.status === 201) ...Another scenario }); ``` Unfortunately, my code did not work and it took some time in order to find out the reason. It seems that _Networking Engine_ does not provide **Response Status** which is weird, because it provides almost all the information. **What did you expect to happen?** I was expecting that **response** will include **response.status**. **What actually happened?** As I mentioned before, my License Wrapping implementation did not work, because I was not able to compare response.status. I don't know the reason why it was excluded from response object.
non_process
networking engine does not provide response http status code have you read the and checked for duplicate open issues yes what version of shaka player are you using can you reproduce the issue with our latest release version yes can you reproduce the issue with the latest code from master yes are you using the demo app or your own custom app irrelevant if custom app can you reproduce the issue using our demo app irrelevant what browser and os are you using irrelevant for embedded devices smart tvs etc what model and firmware version are you using irrelevant what are the manifest and license server uris irrelevant what configuration are you using what is the output of player getconfiguration irrelevant what did you do i was trying to implement license wrapping based on the official tutorial as the result i had the following code js player getnetworkingengine registerresponsefilter function type response if type shaka net networkingengine requesttype license return if response status one scenario if response status another scenario unfortunately my code did not work and it took some time in order to find out the reason it seems that networking engine does not provide response status which is weird because it provides almost all the information what did you expect to happen i was expecting that response will include response status what actually happened as i mentioned before my license wrapping implementation did not work because i was not able to compare response status i don t know the reason why it was excluded from response object
0
8,769
11,886,378,192
IssuesEvent
2020-03-27 21:46:34
MicrosoftDocs/vsts-docs
https://api.github.com/repos/MicrosoftDocs/vsts-docs
closed
Step condition doesn't work
Pri1 devops-cicd-process/tech devops/prod support-request
After https://github.com/Pr0methean/BetterRandom/commit/abd1a4353829ee3725527cd2b4734c686b678597, I still don't get any artifacts after the previous step fails. Why not? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3f151218-9a11-0078-e038-f96198a76143 * Version Independent ID: 09c4d032-62f3-d97c-79d7-6fbfd89910e9 * Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml#feedback) * Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/conditions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Step condition doesn't work - After https://github.com/Pr0methean/BetterRandom/commit/abd1a4353829ee3725527cd2b4734c686b678597, I still don't get any artifacts after the previous step fails. Why not? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3f151218-9a11-0078-e038-f96198a76143 * Version Independent ID: 09c4d032-62f3-d97c-79d7-6fbfd89910e9 * Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml#feedback) * Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/conditions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
step condition doesn t work after i still don t get any artifacts after the previous step fails why not document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
13,810
16,569,500,969
IssuesEvent
2021-05-30 05:05:02
trpo2021/cw-ip-011_keyboardninja
https://api.github.com/repos/trpo2021/cw-ip-011_keyboardninja
opened
merging branches
in process
need to carefully merge the branches of the test2 graphics and main, along the way, correct the tests and makefile
1.0
merging branches - need to carefully merge the branches of the test2 graphics and main, along the way, correct the tests and makefile
process
merging branches need to carefully merge the branches of the graphics and main along the way correct the tests and makefile
1
160,877
25,248,533,241
IssuesEvent
2022-11-15 13:02:38
Sun-Mountain/lettuceMeetApp
https://api.github.com/repos/Sun-Mountain/lettuceMeetApp
closed
Feature - Make nav bar collapsable when viewed on a smaller screen
content-design description-needed
<!-- e.g. Title should be describing the story/feature in one sentences: - As a team user, I want to be able to move a student from one roster to another. - Create Title Case component and use for all Nav Items, Section Titles, and Table Headers --> **Description and related issues -** <!-- Describe the feature here and link or add dependencies to any applicable issues; include background of the problem being solved, policy, regulation, or business justification --> **Does this need to be broken up into smaller issues?** **What is the goal or purpose of this ticket?** **Mockups or Screenshots** ### Acceptance criteria | Given | When | Then | | ----- | ---- | ---- | | | | | ### This task is done when… This could include: - [ ] a list of changes to be made - [ ] all acceptance criteria are met - [ ] the description of what a user might be able to accomplish once this is done - [ ] identify 508 considerations and/or recommendations --- <!-- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. -->
1.0
Feature - Make nav bar collapsable when viewed on a smaller screen - <!-- e.g. Title should be describing the story/feature in one sentences: - As a team user, I want to be able to move a student from one roster to another. - Create Title Case component and use for all Nav Items, Section Titles, and Table Headers --> **Description and related issues -** <!-- Describe the feature here and link or add dependencies to any applicable issues; include background of the problem being solved, policy, regulation, or business justification --> **Does this need to be broken up into smaller issues?** **What is the goal or purpose of this ticket?** **Mockups or Screenshots** ### Acceptance criteria | Given | When | Then | | ----- | ---- | ---- | | | | | ### This task is done when… This could include: - [ ] a list of changes to be made - [ ] all acceptance criteria are met - [ ] the description of what a user might be able to accomplish once this is done - [ ] identify 508 considerations and/or recommendations --- <!-- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. -->
non_process
feature make nav bar collapsable when viewed on a smaller screen e g title should be describing the story feature in one sentences as a team user i want to be able to move a student from one roster to another create title case component and use for all nav items section titles and table headers description and related issues does this need to be broken up into smaller issues what is the goal or purpose of this ticket mockups or screenshots acceptance criteria given when then this task is done when… this could include a list of changes to be made all acceptance criteria are met the description of what a user might be able to accomplish once this is done identify considerations and or recommendations
0
303,226
26,194,127,000
IssuesEvent
2023-01-03 11:52:57
navikt/tiltaksgjennomforing-prosess
https://api.github.com/repos/navikt/tiltaksgjennomforing-prosess
closed
Bygg av testing-dokumentfordeling
deploy dev-fss testing-dokumentfordeling
Kommenter med >/deploy testing-dokumentfordeling for å deploye til dev-fss. Commit: 494723237ef34f1de33270cc3c1b78943f51fb48
1.0
Bygg av testing-dokumentfordeling - Kommenter med >/deploy testing-dokumentfordeling for å deploye til dev-fss. Commit: 494723237ef34f1de33270cc3c1b78943f51fb48
non_process
bygg av testing dokumentfordeling kommenter med deploy testing dokumentfordeling for å deploye til dev fss commit
0
16,043
20,190,730,255
IssuesEvent
2022-02-11 05:01:06
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Create runbook step 2 incorrectly indicates to select Powershell instead of PowerShell Workflow
automation/svc triaged cxp doc-enhancement process-automation/subsvc Pri2
[Enter feedback here] Step 2 of the creating a Workbook indicates to create a PowerShell workbook and should be a PowerShell Workflow --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3632c749-8963-f5ed-55ec-28af005780bd * Version Independent ID: 3ec0f957-e320-7ea7-e5f5-07f543f3c31b * Content: [Tutorial - Create a PowerShell Workflow runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual#feedback) * Content Source: [articles/automation/learn/automation-tutorial-runbook-textual.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **v-ssudhir**
1.0
Create runbook step 2 incorrectly indicates to select Powershell instead of PowerShell Workflow - [Enter feedback here] Step 2 of the creating a Workbook indicates to create a PowerShell workbook and should be a PowerShell Workflow --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3632c749-8963-f5ed-55ec-28af005780bd * Version Independent ID: 3ec0f957-e320-7ea7-e5f5-07f543f3c31b * Content: [Tutorial - Create a PowerShell Workflow runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual#feedback) * Content Source: [articles/automation/learn/automation-tutorial-runbook-textual.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **v-ssudhir**
process
create runbook step incorrectly indicates to select powershell instead of powershell workflow step of the creating a workbook indicates to create a powershell workbook and should be a powershell workflow document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
1
12,995
15,359,205,502
IssuesEvent
2021-03-01 15:36:32
edwardsmarc/CASFRI
https://api.github.com/repos/edwardsmarc/CASFRI
opened
Create a DropConstraints.sql script
blocker enhancement post-translation process translation
So constraints can be droped befere retranslating an inventory (should make translation faster) or before translating a new one which does not fullfil all the constraints.
1.0
Create a DropConstraints.sql script - So constraints can be droped befere retranslating an inventory (should make translation faster) or before translating a new one which does not fullfil all the constraints.
process
create a dropconstraints sql script so constraints can be droped befere retranslating an inventory should make translation faster or before translating a new one which does not fullfil all the constraints
1
12,300
14,856,300,329
IssuesEvent
2021-01-18 13:58:37
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Grouping by required fields should be not have nullable output type
bug/2-confirmed kind/bug process/candidate team/client topic: groupBy
## Problem ```prisma type User { id Int @id name String } ``` ```ts const user = prisma.user.groupBy({ by: ['name'] }) user[0].name // string | null ``` ## Suggested solution If `name` can't be null in the schema, then the name field in the result set can't be null. It should be: ```ts user[0].name // string ``` ## Original Ticket #4893
1.0
Grouping by required fields should be not have nullable output type - ## Problem ```prisma type User { id Int @id name String } ``` ```ts const user = prisma.user.groupBy({ by: ['name'] }) user[0].name // string | null ``` ## Suggested solution If `name` can't be null in the schema, then the name field in the result set can't be null. It should be: ```ts user[0].name // string ``` ## Original Ticket #4893
process
grouping by required fields should be not have nullable output type problem prisma type user id int id name string ts const user prisma user groupby by user name string null suggested solution if name can t be null in the schema then the name field in the result set can t be null it should be ts user name string original ticket
1
11,950
14,713,115,234
IssuesEvent
2021-01-05 09:53:59
nestauk/sg_covid_impact
https://api.github.com/repos/nestauk/sg_covid_impact
closed
Load and process BRES and IDBR data
processing
We use getters to load and process BRES and IDBR data into division level.
1.0
Load and process BRES and IDBR data - We use getters to load and process BRES and IDBR data into division level.
process
load and process bres and idbr data we use getters to load and process bres and idbr data into division level
1
18,259
24,341,438,806
IssuesEvent
2022-10-01 19:05:32
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
closed
Tidy filetypes for datasets
good first issue data processing back end
Filetypes for datasets are getting a bit out of hand. Currently we display them directly as provided by publisher, but inconsistencies are starting to show (e.g. "ZIP" and ".ZIP") This should be a relatively easy solution to apply and it would be quite similar to how we tidy licensing and category information already (see https://github.com/OpenDataScotland/the_od_bods/blob/main/merge_data.py) ** Note that in https://opendata.scot/datasets.json the filetype is the resource [Name]. In https://raw.githubusercontent.com/OpenDataScotland/the_od_bods/main/data/merged_output.csv it is the [FileType] See original conversation in slack: https://opendatascotland.slack.com/archives/C02HEHDL8AY/p1655743400444269
1.0
Tidy filetypes for datasets - Filetypes for datasets are getting a bit out of hand. Currently we display them directly as provided by publisher, but inconsistencies are starting to show (e.g. "ZIP" and ".ZIP") This should be a relatively easy solution to apply and it would be quite similar to how we tidy licensing and category information already (see https://github.com/OpenDataScotland/the_od_bods/blob/main/merge_data.py) ** Note that in https://opendata.scot/datasets.json the filetype is the resource [Name]. In https://raw.githubusercontent.com/OpenDataScotland/the_od_bods/main/data/merged_output.csv it is the [FileType] See original conversation in slack: https://opendatascotland.slack.com/archives/C02HEHDL8AY/p1655743400444269
process
tidy filetypes for datasets filetypes for datasets are getting a bit out of hand currently we display them directly as provided by publisher but inconsistencies are starting to show e g zip and zip this should be a relatively easy solution to apply and it would be quite similar to how we tidy licensing and category information already see note that in the filetype is the resource in it is the see original conversation in slack
1
15,652
19,846,793,876
IssuesEvent
2022-01-21 07:38:47
ooi-data/CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth
https://api.github.com/repos/ooi-data/CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:38:46.523700. ## Details Flow name: `CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:38:46.523700. ## Details Flow name: `CE09OSSM-RID26-01-ADCPTC000-recovered_host-adcp_velocity_earth` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host adcp velocity earth task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
147,373
19,520,315,274
IssuesEvent
2021-12-29 17:10:34
mregen/UA-.NetStandardLibrary
https://api.github.com/repos/mregen/UA-.NetStandardLibrary
closed
CVE-2019-1302 (High) detected in microsoft.netcore.app.2.1.0.nupkg
security vulnerability
## CVE-2019-1302 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>microsoft.netcore.app.2.1.0.nupkg</b></p></summary> <p>A set of .NET API's that are included in the default .NET Core application model. caa7b7e2bad98e56a...</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg</a></p> <p>Path to dependency file: /Tests/Opc.Ua.Security.Certificates.Tests/Opc.Ua.Security.Certificates.Tests.csproj</p> <p>Path to vulnerable library: /microsoft.netcore.app/2.1.0/microsoft.netcore.app.2.1.0.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.netcore.app.2.1.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mregen/UA-.NetStandardLibrary/commit/cc7c7249fb08f768b869a09371e53abf3b2c2047">cc7c7249fb08f768b869a09371e53abf3b2c2047</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An elevation of privilege vulnerability exists when a ASP.NET Core web application, created using vulnerable project templates, fails to properly sanitize web requests, aka 'ASP.NET Core Elevation Of Privilege Vulnerability'. <p>Publish Date: 2019-09-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1302>CVE-2019-1302</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/aspnet/Announcements/issues/384">https://github.com/aspnet/Announcements/issues/384</a></p> <p>Release Date: 2019-09-12</p> <p>Fix Resolution: Microsoft.AspNetCore.SpaServices - 2.2.1,2.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-1302 (High) detected in microsoft.netcore.app.2.1.0.nupkg - ## CVE-2019-1302 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>microsoft.netcore.app.2.1.0.nupkg</b></p></summary> <p>A set of .NET API's that are included in the default .NET Core application model. caa7b7e2bad98e56a...</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.1.0.nupkg</a></p> <p>Path to dependency file: /Tests/Opc.Ua.Security.Certificates.Tests/Opc.Ua.Security.Certificates.Tests.csproj</p> <p>Path to vulnerable library: /microsoft.netcore.app/2.1.0/microsoft.netcore.app.2.1.0.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.netcore.app.2.1.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mregen/UA-.NetStandardLibrary/commit/cc7c7249fb08f768b869a09371e53abf3b2c2047">cc7c7249fb08f768b869a09371e53abf3b2c2047</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An elevation of privilege vulnerability exists when a ASP.NET Core web application, created using vulnerable project templates, fails to properly sanitize web requests, aka 'ASP.NET Core Elevation Of Privilege Vulnerability'. <p>Publish Date: 2019-09-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1302>CVE-2019-1302</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/aspnet/Announcements/issues/384">https://github.com/aspnet/Announcements/issues/384</a></p> <p>Release Date: 2019-09-12</p> <p>Fix Resolution: Microsoft.AspNetCore.SpaServices - 2.2.1,2.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in microsoft netcore app nupkg cve high severity vulnerability vulnerable library microsoft netcore app nupkg a set of net api s that are included in the default net core application model library home page a href path to dependency file tests opc ua security certificates tests opc ua security certificates tests csproj path to vulnerable library microsoft netcore app microsoft netcore app nupkg dependency hierarchy x microsoft netcore app nupkg vulnerable library found in head commit a href vulnerability details an elevation of privilege vulnerability exists when a asp net core web application created using vulnerable project templates fails to properly sanitize web requests aka asp net core elevation of privilege vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore spaservices step up your open source security game with whitesource
0
594,453
18,046,020,883
IssuesEvent
2021-09-18 22:56:59
python/mypy
https://api.github.com/repos/python/mypy
closed
Crash in proper_plugin.py when one argument passed to `isinstance`
crash priority-2-low
**Crash Report** When using the `misc/proper_plugin.py` plugin, mypy crashes when only one argument is passed to `isinstance` **Traceback** ``` test.py:1: error: Too few arguments for "isinstance" test.py:1: error: INTERNAL ERROR -- Please try using mypy master on Github: https://mypy.readthedocs.io/en/stable/common_issues.html#using-a-development-mypy-build Please report a bug at https://github.com/python/mypy/issues version: 0.910 Traceback (most recent call last): File "mypy/checkexpr.py", line 3905, in accept File "mypy/checkexpr.py", line 271, in visit_call_expr File "mypy/checkexpr.py", line 353, in visit_call_expr_inner File "mypy/checkexpr.py", line 858, in check_call_expr_with_callee_type File "mypy/checkexpr.py", line 917, in check_call File "mypy/checkexpr.py", line 1029, in check_callable_call File "mypy/checkexpr.py", line 729, in apply_function_plugin File "/mnt/e/Pranav/Github/mypy/misc/proper_plugin.py", line 37, in isinstance_proper_hook right = get_proper_type(ctx.arg_types[1][0]) IndexError: list index out of range test.py:1: : note: use --pdb to drop into pdb ``` **To Reproduce** test.py: ```py isinstance(5) ``` Run mypy with `mypy test.py --show-traceback` (or just `mypy test.py`). **Your Environment** - Mypy version used: 0.910 - Mypy command-line flags: None necessary - Mypy configuration options from `mypy.ini` (and other config files): mypy.ini: ```ini [mypy] # Change this path if you're not at the root of the mypy repo plugins = misc/proper_plugin.py ``` - Python version used: 3.8.5 - Operating system and version: Ubuntu 20.04.2 LTS on WSL
1.0
Crash in proper_plugin.py when one argument passed to `isinstance` - **Crash Report** When using the `misc/proper_plugin.py` plugin, mypy crashes when only one argument is passed to `isinstance` **Traceback** ``` test.py:1: error: Too few arguments for "isinstance" test.py:1: error: INTERNAL ERROR -- Please try using mypy master on Github: https://mypy.readthedocs.io/en/stable/common_issues.html#using-a-development-mypy-build Please report a bug at https://github.com/python/mypy/issues version: 0.910 Traceback (most recent call last): File "mypy/checkexpr.py", line 3905, in accept File "mypy/checkexpr.py", line 271, in visit_call_expr File "mypy/checkexpr.py", line 353, in visit_call_expr_inner File "mypy/checkexpr.py", line 858, in check_call_expr_with_callee_type File "mypy/checkexpr.py", line 917, in check_call File "mypy/checkexpr.py", line 1029, in check_callable_call File "mypy/checkexpr.py", line 729, in apply_function_plugin File "/mnt/e/Pranav/Github/mypy/misc/proper_plugin.py", line 37, in isinstance_proper_hook right = get_proper_type(ctx.arg_types[1][0]) IndexError: list index out of range test.py:1: : note: use --pdb to drop into pdb ``` **To Reproduce** test.py: ```py isinstance(5) ``` Run mypy with `mypy test.py --show-traceback` (or just `mypy test.py`). **Your Environment** - Mypy version used: 0.910 - Mypy command-line flags: None necessary - Mypy configuration options from `mypy.ini` (and other config files): mypy.ini: ```ini [mypy] # Change this path if you're not at the root of the mypy repo plugins = misc/proper_plugin.py ``` - Python version used: 3.8.5 - Operating system and version: Ubuntu 20.04.2 LTS on WSL
non_process
crash in proper plugin py when one argument passed to isinstance crash report when using the misc proper plugin py plugin mypy crashes when only one argument is passed to isinstance traceback test py error too few arguments for isinstance test py error internal error please try using mypy master on github please report a bug at version traceback most recent call last file mypy checkexpr py line in accept file mypy checkexpr py line in visit call expr file mypy checkexpr py line in visit call expr inner file mypy checkexpr py line in check call expr with callee type file mypy checkexpr py line in check call file mypy checkexpr py line in check callable call file mypy checkexpr py line in apply function plugin file mnt e pranav github mypy misc proper plugin py line in isinstance proper hook right get proper type ctx arg types indexerror list index out of range test py note use pdb to drop into pdb to reproduce test py py isinstance run mypy with mypy test py show traceback or just mypy test py your environment mypy version used mypy command line flags none necessary mypy configuration options from mypy ini and other config files mypy ini ini change this path if you re not at the root of the mypy repo plugins misc proper plugin py python version used operating system and version ubuntu lts on wsl
0
8,538
11,713,944,699
IssuesEvent
2020-03-09 11:18:49
kazuwjnlab/cvpaper
https://api.github.com/repos/kazuwjnlab/cvpaper
opened
[cvpaper] CVPR2019 #71 Label Efficient Semi-Supervised Learning via Graph Filtering
graph graph convolutional neural network graph signal processing semi-supervised learning
## \#71 [Label Efficient Semi-Supervised Learning via Graph Filtering](http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Label_Efficient_Semi-Supervised_Learning_via_Graph_Filtering_CVPR_2019_paper.pdf) Qimai Li, Xiao-Ming Wu, Han Liu, Xiaotong Zhang, Zhichao Guan ### どんな論文か? グラフベース半教師あり学習は、ラベル間の接続情報を使用できる点で他の半教師あり学習の手法に比べて有利だが、ラベル伝播の手法が古典的なものであったり、NN系の手法を用いるためにラベル付きデータを大量に必要とするためラベル効率が悪い問題があった。この論文では、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングで特徴抽出を行う。さらに、フィルターの強度でラベル効率を操作することも可能となった。 ### 新規性 グラフベース半教師あり学習において、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングによって特徴抽出を行っている点 ### 結果 半教師あり分類、回帰、ゼロショット画像認識でほぼほぼ SoTA ### その他(なぜ通ったか?等) graph signal processing の問題に化けさせる論文は読んだことがないので面白い切り口かと。あと、結果がよい。 ![thumb]() graph, graph convolutional neural network, semi-supervised learning, graph signal processing,
1.0
[cvpaper] CVPR2019 #71 Label Efficient Semi-Supervised Learning via Graph Filtering - ## \#71 [Label Efficient Semi-Supervised Learning via Graph Filtering](http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Label_Efficient_Semi-Supervised_Learning_via_Graph_Filtering_CVPR_2019_paper.pdf) Qimai Li, Xiao-Ming Wu, Han Liu, Xiaotong Zhang, Zhichao Guan ### どんな論文か? グラフベース半教師あり学習は、ラベル間の接続情報を使用できる点で他の半教師あり学習の手法に比べて有利だが、ラベル伝播の手法が古典的なものであったり、NN系の手法を用いるためにラベル付きデータを大量に必要とするためラベル効率が悪い問題があった。この論文では、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングで特徴抽出を行う。さらに、フィルターの強度でラベル効率を操作することも可能となった。 ### 新規性 グラフベース半教師あり学習において、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングによって特徴抽出を行っている点 ### 結果 半教師あり分類、回帰、ゼロショット画像認識でほぼほぼ SoTA ### その他(なぜ通ったか?等) graph signal processing の問題に化けさせる論文は読んだことがないので面白い切り口かと。あと、結果がよい。 ![thumb]() graph, graph convolutional neural network, semi-supervised learning, graph signal processing,
process
label efficient semi supervised learning via graph filtering qimai li xiao ming wu han liu xiaotong zhang zhichao guan どんな論文か? グラフベース半教師あり学習は、ラベル間の接続情報を使用できる点で他の半教師あり学習の手法に比べて有利だが、ラベル伝播の手法が古典的なものであったり、nn系の手法を用いるためにラベル付きデータを大量に必要とするためラベル効率が悪い問題があった。この論文では、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングで特徴抽出を行う。さらに、フィルターの強度でラベル効率を操作することも可能となった。 新規性 グラフベース半教師あり学習において、ラベル伝播をグラフ上での信号伝播と捉え、グラフフィルタリングによって特徴抽出を行っている点 結果 半教師あり分類、回帰、ゼロショット画像認識でほぼほぼ sota その他(なぜ通ったか?等) graph signal processing の問題に化けさせる論文は読んだことがないので面白い切り口かと。あと、結果がよい。 graph graph convolutional neural network semi supervised learning graph signal processing
1
7,051
10,210,693,024
IssuesEvent
2019-08-14 15:17:27
pelias/pelias
https://api.github.com/repos/pelias/pelias
closed
Support multiple boundary.country parameters in a request
enhancement good first issue help wanted processed
#### Here's what I did :innocent: This came in through support (see desk 882). The user is looking for a way to filter by multiple countries with autocomplete, although this applies to other endpoints as well. Example query: https://mapzen.com/search/explorer/?query=search&text=ymca&boundary.country=GBR%2CIRL&focus.point.lat=-33.856680&focus.point.lon=151.215281 --- #### Here's what I got :scream_cat: `[ERROR] GBR,IRL is not a valid ISO2/ISO3 country code` --- #### Here's what I was expecting :sparkles: The ability to use something like `boundary.country=GBR,IRL` or `boundary.country=GB&boundary.country=DE` with separate query parameters to filter by multiple countries, similar to`layers` and `sources`. --- #### Here's what I think could be improved :trophy: @dianashk confirmed that using multiple countries is not currently supported. If the countries are close enough, then you could use `boundary.rect` param, or do some filtering on your own based on the records.
1.0
Support multiple boundary.country parameters in a request - #### Here's what I did :innocent: This came in through support (see desk 882). The user is looking for a way to filter by multiple countries with autocomplete, although this applies to other endpoints as well. Example query: https://mapzen.com/search/explorer/?query=search&text=ymca&boundary.country=GBR%2CIRL&focus.point.lat=-33.856680&focus.point.lon=151.215281 --- #### Here's what I got :scream_cat: `[ERROR] GBR,IRL is not a valid ISO2/ISO3 country code` --- #### Here's what I was expecting :sparkles: The ability to use something like `boundary.country=GBR,IRL` or `boundary.country=GB&boundary.country=DE` with separate query parameters to filter by multiple countries, similar to`layers` and `sources`. --- #### Here's what I think could be improved :trophy: @dianashk confirmed that using multiple countries is not currently supported. If the countries are close enough, then you could use `boundary.rect` param, or do some filtering on your own based on the records.
process
support multiple boundary country parameters in a request here s what i did innocent this came in through support see desk the user is looking for a way to filter by multiple countries with autocomplete although this applies to other endpoints as well example query here s what i got scream cat gbr irl is not a valid country code here s what i was expecting sparkles the ability to use something like boundary country gbr irl or boundary country gb boundary country de with separate query parameters to filter by multiple countries similar to layers and sources here s what i think could be improved trophy dianashk confirmed that using multiple countries is not currently supported if the countries are close enough then you could use boundary rect param or do some filtering on your own based on the records
1
11,399
14,235,002,051
IssuesEvent
2020-11-18 14:18:44
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
Icons for PH calcs
Calculator Process Heating
Let me know if they still have the white background and I'll send them in slack Flue Gas ![flueGas.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/664880b7-b61b-4a07-b33b-2925abb112de) Wall ![wall.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/8a3c0d54-ce89-4a0e-8de3-ce22ddd78feb) Hot Air Leak ![hotAirLeak.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/b3fcf34d-fc93-4a43-b8bd-211e78678f88) Fixture ![fixture.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/d80e41f0-8091-41bb-bedf-6364d64df24c) Opening ![opening.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/1c792b2b-0b6f-48ed-ba8a-fd9ca9ddf2fa) Cooling ![cooling.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/eed7e717-671a-455a-be8e-c64d65960f61) Atmosphere ![atmosphere.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/14c69ea0-e357-440e-911b-09e689743c82) Charge Materials To Do
1.0
Icons for PH calcs - Let me know if they still have the white background and I'll send them in slack Flue Gas ![flueGas.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/664880b7-b61b-4a07-b33b-2925abb112de) Wall ![wall.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/8a3c0d54-ce89-4a0e-8de3-ce22ddd78feb) Hot Air Leak ![hotAirLeak.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/b3fcf34d-fc93-4a43-b8bd-211e78678f88) Fixture ![fixture.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/d80e41f0-8091-41bb-bedf-6364d64df24c) Opening ![opening.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/1c792b2b-0b6f-48ed-ba8a-fd9ca9ddf2fa) Cooling ![cooling.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/eed7e717-671a-455a-be8e-c64d65960f61) Atmosphere ![atmosphere.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/14c69ea0-e357-440e-911b-09e689743c82) Charge Materials To Do
process
icons for ph calcs let me know if they still have the white background and i ll send them in slack flue gas wall hot air leak fixture opening cooling atmosphere charge materials to do
1
536,257
15,706,787,326
IssuesEvent
2021-03-26 17:54:54
yalla-coop/chiltern-music-therapy
https://api.github.com/repos/yalla-coop/chiltern-music-therapy
opened
I see a dashboard as a client
back-end front-end priority-3
__Wireframe link__ https://www.figma.com/file/CcYmhfnXreAPxlfyEmGsAH/Chiltern-Music-Therapy?node-id=469%3A34091 --- ### Acceptance Criteria: _REMEMBER THAT WHOEVER WORKS ON THIS ISSUE MUST TICK OFF ALL THE POINTS IN THIS LIST UNLESS THERE IS CLEAR AGREEMENT IN THE COMMENTS TO SAY OTHERWISE. **DO NOT REVIEW A PR INVOLVING THIS ISSUE UNLESS THIS HAS BEEN DONE**_ - [ ] Set up page in line with wireframes - [ ] Api call and database query to fetch - [ ] Initials of the patient (i.e. first character of each first and last name) - [ ] Name of therapist - [ ] The latest programme to be created - [ ] To confirm, view more is a link that goes to the My Programmes Page
1.0
I see a dashboard as a client - __Wireframe link__ https://www.figma.com/file/CcYmhfnXreAPxlfyEmGsAH/Chiltern-Music-Therapy?node-id=469%3A34091 --- ### Acceptance Criteria: _REMEMBER THAT WHOEVER WORKS ON THIS ISSUE MUST TICK OFF ALL THE POINTS IN THIS LIST UNLESS THERE IS CLEAR AGREEMENT IN THE COMMENTS TO SAY OTHERWISE. **DO NOT REVIEW A PR INVOLVING THIS ISSUE UNLESS THIS HAS BEEN DONE**_ - [ ] Set up page in line with wireframes - [ ] Api call and database query to fetch - [ ] Initials of the patient (i.e. first character of each first and last name) - [ ] Name of therapist - [ ] The latest programme to be created - [ ] To confirm, view more is a link that goes to the My Programmes Page
non_process
i see a dashboard as a client wireframe link acceptance criteria remember that whoever works on this issue must tick off all the points in this list unless there is clear agreement in the comments to say otherwise do not review a pr involving this issue unless this has been done set up page in line with wireframes api call and database query to fetch initials of the patient i e first character of each first and last name name of therapist the latest programme to be created to confirm view more is a link that goes to the my programmes page
0
195,273
22,300,743,141
IssuesEvent
2022-06-13 08:29:31
Trinadh465/gson_parent-2.8.8
https://api.github.com/repos/Trinadh465/gson_parent-2.8.8
opened
CVE-2020-15250 (Medium) detected in junit-4.12.jar
security vulnerability
## CVE-2020-15250 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junit-4.12.jar</b></p></summary> <p>JUnit is a unit testing framework for Java, created by Erich Gamma and Kent Beck.</p> <p>Library home page: <a href="http://junit.org">http://junit.org</a></p> <p>Path to dependency file: /gson/build.gradle</p> <p>Path to vulnerable library: /hes/modules-2/files-2.1/junit/junit/4.12/2973d150c0dc1fefe998f834810d68f278ea58ec/junit-4.12.jar</p> <p> Dependency Hierarchy: - :x: **junit-4.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/gson_parent-2.8.8/commit/52f0dd349a1379714ba0cc09a4f1315f471c8ca8">52f0dd349a1379714ba0cc09a4f1315f471c8ca8</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In JUnit4 from version 4.7 and before 4.13.1, the test rule TemporaryFolder contains a local information disclosure vulnerability. On Unix like systems, the system's temporary directory is shared between all users on that system. Because of this, when files and directories are written into this directory they are, by default, readable by other users on that same system. This vulnerability does not allow other users to overwrite the contents of these directories or files. This is purely an information disclosure vulnerability. This vulnerability impacts you if the JUnit tests write sensitive information, like API keys or passwords, into the temporary folder, and the JUnit tests execute in an environment where the OS has other untrusted users. Because certain JDK file system APIs were only added in JDK 1.7, this this fix is dependent upon the version of the JDK you are using. For Java 1.7 and higher users: this vulnerability is fixed in 4.13.1. For Java 1.6 and lower users: no patch is available, you must use the workaround below. If you are unable to patch, or are stuck running on Java 1.6, specifying the `java.io.tmpdir` system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability. For more information, including an example of vulnerable code, see the referenced GitHub Security Advisory. <p>Publish Date: 2020-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15250>CVE-2020-15250</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp">https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp</a></p> <p>Release Date: 2020-10-12</p> <p>Fix Resolution: 4.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15250 (Medium) detected in junit-4.12.jar - ## CVE-2020-15250 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junit-4.12.jar</b></p></summary> <p>JUnit is a unit testing framework for Java, created by Erich Gamma and Kent Beck.</p> <p>Library home page: <a href="http://junit.org">http://junit.org</a></p> <p>Path to dependency file: /gson/build.gradle</p> <p>Path to vulnerable library: /hes/modules-2/files-2.1/junit/junit/4.12/2973d150c0dc1fefe998f834810d68f278ea58ec/junit-4.12.jar</p> <p> Dependency Hierarchy: - :x: **junit-4.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/gson_parent-2.8.8/commit/52f0dd349a1379714ba0cc09a4f1315f471c8ca8">52f0dd349a1379714ba0cc09a4f1315f471c8ca8</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In JUnit4 from version 4.7 and before 4.13.1, the test rule TemporaryFolder contains a local information disclosure vulnerability. On Unix like systems, the system's temporary directory is shared between all users on that system. Because of this, when files and directories are written into this directory they are, by default, readable by other users on that same system. This vulnerability does not allow other users to overwrite the contents of these directories or files. This is purely an information disclosure vulnerability. This vulnerability impacts you if the JUnit tests write sensitive information, like API keys or passwords, into the temporary folder, and the JUnit tests execute in an environment where the OS has other untrusted users. Because certain JDK file system APIs were only added in JDK 1.7, this this fix is dependent upon the version of the JDK you are using. For Java 1.7 and higher users: this vulnerability is fixed in 4.13.1. For Java 1.6 and lower users: no patch is available, you must use the workaround below. If you are unable to patch, or are stuck running on Java 1.6, specifying the `java.io.tmpdir` system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability. For more information, including an example of vulnerable code, see the referenced GitHub Security Advisory. <p>Publish Date: 2020-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15250>CVE-2020-15250</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp">https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp</a></p> <p>Release Date: 2020-10-12</p> <p>Fix Resolution: 4.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in junit jar cve medium severity vulnerability vulnerable library junit jar junit is a unit testing framework for java created by erich gamma and kent beck library home page a href path to dependency file gson build gradle path to vulnerable library hes modules files junit junit junit jar dependency hierarchy x junit jar vulnerable library found in head commit a href found in base branch master vulnerability details in from version and before the test rule temporaryfolder contains a local information disclosure vulnerability on unix like systems the system s temporary directory is shared between all users on that system because of this when files and directories are written into this directory they are by default readable by other users on that same system this vulnerability does not allow other users to overwrite the contents of these directories or files this is purely an information disclosure vulnerability this vulnerability impacts you if the junit tests write sensitive information like api keys or passwords into the temporary folder and the junit tests execute in an environment where the os has other untrusted users because certain jdk file system apis were only added in jdk this this fix is dependent upon the version of the jdk you are using for java and higher users this vulnerability is fixed in for java and lower users no patch is available you must use the workaround below if you are unable to patch or are stuck running on java specifying the java io tmpdir system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability for more information including an example of vulnerable code see the referenced github security advisory publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
4,978
7,808,418,722
IssuesEvent
2018-06-11 20:07:25
GoogleCloudPlatform/google-cloud-cpp
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
closed
Run integration tests against emulator
storage testing type: process
To get code coverage statistics that make any sense we need to run the integration tests against some kind of emulator. Maybe we can reuse the python stuff we did for httpbin and add other handlers for the emulator. All we need is to keep buckets and objects in memory.
1.0
Run integration tests against emulator - To get code coverage statistics that make any sense we need to run the integration tests against some kind of emulator. Maybe we can reuse the python stuff we did for httpbin and add other handlers for the emulator. All we need is to keep buckets and objects in memory.
process
run integration tests against emulator to get code coverage statistics that make any sense we need to run the integration tests against some kind of emulator maybe we can reuse the python stuff we did for httpbin and add other handlers for the emulator all we need is to keep buckets and objects in memory
1
640,728
20,797,615,564
IssuesEvent
2022-03-17 10:49:26
slsdetectorgroup/slsDetectorPackage
https://api.github.com/repos/slsdetectorgroup/slsDetectorPackage
closed
Receiver: disabled ports write files in 10g
action - Bug priority - High status - resolved
<!-- Preview changes before submitting --> <!-- Please fill out everything with an *, as this report will be discarded otherwise --> <!-- This is a comment, the syntax is a bit different from c++ or bash --> ##### *Distribution: <!-- RHEL7, RHEL6, Fedora, etc --> ##### *Detector type: <!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard --> Receiverv ##### *Software Package Version: <!-- developer, 4.2.0, 4.1.1, etc --> ##### Priority: <!-- Super Low, Low, Medium, High, Super High --> High ##### *Describe the bug <!-- A clear and concise description of what the bug is --> Disabled ports write files ##### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ##### To Reproduce <!-- Steps to reproduce the behavior: --> <!-- 1. Go to '...' --> <!-- 2. Click on '....' --> <!-- 3. Scroll down to '....' --> <!-- 4. See error --> ##### Screenshots <!-- If applicable, add screenshots to help explain your problem. --> ##### Additional context <!-- Add any other context about the problem here. -->
1.0
Receiver: disabled ports write files in 10g - <!-- Preview changes before submitting --> <!-- Please fill out everything with an *, as this report will be discarded otherwise --> <!-- This is a comment, the syntax is a bit different from c++ or bash --> ##### *Distribution: <!-- RHEL7, RHEL6, Fedora, etc --> ##### *Detector type: <!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard --> Receiverv ##### *Software Package Version: <!-- developer, 4.2.0, 4.1.1, etc --> ##### Priority: <!-- Super Low, Low, Medium, High, Super High --> High ##### *Describe the bug <!-- A clear and concise description of what the bug is --> Disabled ports write files ##### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ##### To Reproduce <!-- Steps to reproduce the behavior: --> <!-- 1. Go to '...' --> <!-- 2. Click on '....' --> <!-- 3. Scroll down to '....' --> <!-- 4. See error --> ##### Screenshots <!-- If applicable, add screenshots to help explain your problem. --> ##### Additional context <!-- Add any other context about the problem here. -->
non_process
receiver disabled ports write files in distribution detector type receiverv software package version priority high describe the bug disabled ports write files expected behavior to reproduce screenshots additional context
0
2,332
5,142,636,883
IssuesEvent
2017-01-12 13:55:36
jimbrown75/Permit-Vision-Enhancements
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
opened
Include Template Verifier and Authoriser signatures in permit created from Template
bug Medium Priority Should Fix Verified by PTW Process Lead
When creating a permit from a template the template verifier and authoriser should be included on the Permit signatures (electronic and printed)
1.0
Include Template Verifier and Authoriser signatures in permit created from Template - When creating a permit from a template the template verifier and authoriser should be included on the Permit signatures (electronic and printed)
process
include template verifier and authoriser signatures in permit created from template when creating a permit from a template the template verifier and authoriser should be included on the permit signatures electronic and printed
1
129,104
10,563,213,043
IssuesEvent
2019-10-04 20:21:24
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Kuberntes 1.15 as officially supported on Rancher 2.2.x
[zube]: To Test team/ca
Currently k8s 1.15 is experimental on 2.2.x. We have to enable an official support for it, and drop support for k8s 1.12 with that change.
1.0
Kuberntes 1.15 as officially supported on Rancher 2.2.x - Currently k8s 1.15 is experimental on 2.2.x. We have to enable an official support for it, and drop support for k8s 1.12 with that change.
non_process
kuberntes as officially supported on rancher x currently is experimental on x we have to enable an official support for it and drop support for with that change
0
19,155
11,156,698,344
IssuesEvent
2019-12-25 08:37:23
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
closed
kubesphere安装配置的外接es,日志系统未开启,安装后部分pod无法启动
area/logging area/microservice
安装环境: 4核4G * 3 centos7 没有开启日志系统,配置如下: ![image](https://user-images.githubusercontent.com/17722518/70675825-3ade3e80-1cc5-11ea-82d2-ddae94b0551e.png) 安装完成后,其他组件正常,现在istio有两个组件异常 ![image](https://user-images.githubusercontent.com/17722518/70675921-7da01680-1cc5-11ea-86b9-245c82295e2e.png) 报错信息如下: ``` [root@master2 ~]# kubectl logs jaeger-collector-8698b58b55-gh7h9 -n istio-system 2019/12/12 01:16:37 maxprocs: Leaving GOMAXPROCS=4: CPU quota undefined {"level":"info","ts":1576113397.8572783,"caller":"flags/service.go:115","msg":"Mounting metrics handler on admin server","route":"/metrics"} {"level":"info","ts":1576113397.8575335,"caller":"flags/admin.go:108","msg":"Mounting health check on admin server","route":"/"} {"level":"info","ts":1576113397.8576038,"caller":"flags/admin.go:114","msg":"Starting admin HTTP server","http-port":14269} {"level":"info","ts":1576113397.857621,"caller":"flags/admin.go:100","msg":"Admin server started","http-port":14269,"health-status":"unavailable"} {"level":"fatal","ts":1576113402.8768172,"caller":"collector/main.go:89","msg":"Failed to init storage factory","error":"failed to create primary Elasticsearch client: health check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host: no Elasticsearch node available","errorVerbose":"no Elasticsearch node available\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.init.ializers\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:88\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:188\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nhealth check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.(*Client).startupHealthcheck\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:1116\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:244\ngithub.com/jaegertracing/jaeger/pkg/es/config.(*Configuration).NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/pkg/es/config/config.go:100\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:80\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nfailed to create primary Elasticsearch client\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:82\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337","stacktrace":"main.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:89\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200"} Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Scheduled 3m3s default-scheduler Successfully assigned istio-system/jaeger-query-7f9c7c84c-4dkcc to node4 Normal Pulled 2m59s kubelet, node4 Container image "jaegertracing/jaeger-agent:1.13" already present on machine Normal Created 2m59s kubelet, node4 Created container jaeger-agent Normal Started 2m58s kubelet, node4 Started container jaeger-agent Warning Unhealthy 2m57s kubelet, node4 Readiness probe failed: HTTP probe failed with statuscode: 503 Normal Started 2m5s (x4 over 2m59s) kubelet, node4 Started container jaeger-query Warning BackOff 82s (x7 over 2m47s) kubelet, node4 Back-off restarting failed container Normal Pulled 69s (x5 over 3m1s) kubelet, node4 Container image "jaegertracing/jaeger-query:1.13" already present on machine Normal Created 69s (x5 over 3m) kubelet, node4 Created container jaeger-query ``` 感觉是es外接配置异常,麻烦看一下配置
1.0
kubesphere安装配置的外接es,日志系统未开启,安装后部分pod无法启动 - 安装环境: 4核4G * 3 centos7 没有开启日志系统,配置如下: ![image](https://user-images.githubusercontent.com/17722518/70675825-3ade3e80-1cc5-11ea-82d2-ddae94b0551e.png) 安装完成后,其他组件正常,现在istio有两个组件异常 ![image](https://user-images.githubusercontent.com/17722518/70675921-7da01680-1cc5-11ea-86b9-245c82295e2e.png) 报错信息如下: ``` [root@master2 ~]# kubectl logs jaeger-collector-8698b58b55-gh7h9 -n istio-system 2019/12/12 01:16:37 maxprocs: Leaving GOMAXPROCS=4: CPU quota undefined {"level":"info","ts":1576113397.8572783,"caller":"flags/service.go:115","msg":"Mounting metrics handler on admin server","route":"/metrics"} {"level":"info","ts":1576113397.8575335,"caller":"flags/admin.go:108","msg":"Mounting health check on admin server","route":"/"} {"level":"info","ts":1576113397.8576038,"caller":"flags/admin.go:114","msg":"Starting admin HTTP server","http-port":14269} {"level":"info","ts":1576113397.857621,"caller":"flags/admin.go:100","msg":"Admin server started","http-port":14269,"health-status":"unavailable"} {"level":"fatal","ts":1576113402.8768172,"caller":"collector/main.go:89","msg":"Failed to init storage factory","error":"failed to create primary Elasticsearch client: health check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host: no Elasticsearch node available","errorVerbose":"no Elasticsearch node available\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.init.ializers\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:88\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:188\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nhealth check timeout: Head http://elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local:9200: dial tcp: lookup elasticsearch-logging-data.kubesphere-logging-system.svc.cluster.local on 169.254.25.10:53: no such host\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.(*Client).startupHealthcheck\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:1116\ngithub.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic%2ev5.NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/gopkg.in/olivere/elastic.v5/client.go:244\ngithub.com/jaegertracing/jaeger/pkg/es/config.(*Configuration).NewClient\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/pkg/es/config/config.go:100\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:80\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337\nfailed to create primary Elasticsearch client\ngithub.com/jaegertracing/jaeger/plugin/storage/es.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/es/factory.go:82\ngithub.com/jaegertracing/jaeger/plugin/storage.(*Factory).Initialize\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/plugin/storage/factory.go:107\nmain.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:88\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200\nruntime.goexit\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/asm_amd64.s:1337","stacktrace":"main.main.func1\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:89\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:762\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).ExecuteC\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:852\ngithub.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra.(*Command).Execute\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/vendor/github.com/spf13/cobra/command.go:800\nmain.main\n\t/home/travis/gopath/src/github.com/jaegertracing/jaeger/cmd/collector/main.go:180\nruntime.main\n\t/home/travis/.gimme/versions/go1.12.1.linux.amd64/src/runtime/proc.go:200"} Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Scheduled 3m3s default-scheduler Successfully assigned istio-system/jaeger-query-7f9c7c84c-4dkcc to node4 Normal Pulled 2m59s kubelet, node4 Container image "jaegertracing/jaeger-agent:1.13" already present on machine Normal Created 2m59s kubelet, node4 Created container jaeger-agent Normal Started 2m58s kubelet, node4 Started container jaeger-agent Warning Unhealthy 2m57s kubelet, node4 Readiness probe failed: HTTP probe failed with statuscode: 503 Normal Started 2m5s (x4 over 2m59s) kubelet, node4 Started container jaeger-query Warning BackOff 82s (x7 over 2m47s) kubelet, node4 Back-off restarting failed container Normal Pulled 69s (x5 over 3m1s) kubelet, node4 Container image "jaegertracing/jaeger-query:1.13" already present on machine Normal Created 69s (x5 over 3m) kubelet, node4 Created container jaeger-query ``` 感觉是es外接配置异常,麻烦看一下配置
non_process
kubesphere安装配置的外接es,日志系统未开启,安装后部分pod无法启动 安装环境: 没有开启日志系统,配置如下: 安装完成后,其他组件正常,现在istio有两个组件异常 报错信息如下: kubectl logs jaeger collector n istio system maxprocs leaving gomaxprocs cpu quota undefined level info ts caller flags service go msg mounting metrics handler on admin server route metrics level info ts caller flags admin go msg mounting health check on admin server route level info ts caller flags admin go msg starting admin http server http port level info ts caller flags admin go msg admin server started http port health status unavailable level fatal ts caller collector main go msg failed to init storage factory error failed to create primary elasticsearch client health check timeout head dial tcp lookup elasticsearch logging data kubesphere logging system svc cluster local on no such host no elasticsearch node available errorverbose no elasticsearch node available ngithub com jaegertracing jaeger vendor gopkg in olivere elastic init ializers n t home travis gopath src github com jaegertracing jaeger vendor gopkg in olivere elastic client go nruntime main n t home travis gimme versions linux src runtime proc go nruntime goexit n t home travis gimme versions linux src runtime asm s nhealth check timeout head dial tcp lookup elasticsearch logging data kubesphere logging system svc cluster local on no such host ngithub com jaegertracing jaeger vendor gopkg in olivere elastic client startuphealthcheck n t home travis gopath src github com jaegertracing jaeger vendor gopkg in olivere elastic client go ngithub com jaegertracing jaeger vendor gopkg in olivere elastic newclient n t home travis gopath src github com jaegertracing jaeger vendor gopkg in olivere elastic client go ngithub com jaegertracing jaeger pkg es config configuration newclient n t home travis gopath src github com jaegertracing jaeger pkg es config config go ngithub com jaegertracing jaeger plugin storage es factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage es factory go ngithub com jaegertracing jaeger plugin storage factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage factory go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command executec n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go nruntime main n t home travis gimme versions linux src runtime proc go nruntime goexit n t home travis gimme versions linux src runtime asm s nfailed to create primary elasticsearch client ngithub com jaegertracing jaeger plugin storage es factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage es factory go ngithub com jaegertracing jaeger plugin storage factory initialize n t home travis gopath src github com jaegertracing jaeger plugin storage factory go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command executec n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go nruntime main n t home travis gimme versions linux src runtime proc go nruntime goexit n t home travis gimme versions linux src runtime asm s stacktrace main main n t home travis gopath src github com jaegertracing jaeger cmd collector main go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command executec n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go ngithub com jaegertracing jaeger vendor github com cobra command execute n t home travis gopath src github com jaegertracing jaeger vendor github com cobra command go nmain main n t home travis gopath src github com jaegertracing jaeger cmd collector main go nruntime main n t home travis gimme versions linux src runtime proc go events type reason age from message normal scheduled default scheduler successfully assigned istio system jaeger query to normal pulled kubelet container image jaegertracing jaeger agent already present on machine normal created kubelet created container jaeger agent normal started kubelet started container jaeger agent warning unhealthy kubelet readiness probe failed http probe failed with statuscode normal started over kubelet started container jaeger query warning backoff over kubelet back off restarting failed container normal pulled over kubelet container image jaegertracing jaeger query already present on machine normal created over kubelet created container jaeger query 感觉是es外接配置异常,麻烦看一下配置
0
9,049
12,130,108,057
IssuesEvent
2020-04-23 00:30:41
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt
priority: p2 remove-gcp-devrel-py-tools type: process
remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt
1.0
remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/django/requirements-test.txt
process
remove gcp devrel py tools from appengine standard django requirements test txt remove gcp devrel py tools from appengine standard django requirements test txt
1
20,596
27,264,367,261
IssuesEvent
2023-02-22 16:55:50
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
closed
"https://www.googleapis.com/auth/bigquery" scope is unecessary
api: bigquery type: process
https://github.com/googleapis/python-bigquery/blob/e1aa9218ad22f85c9a6cab8b61d013779376a582/google/cloud/bigquery/client.py#L229 `"https://www.googleapis.com/auth/cloud-platform"` is a superset of `"https://www.googleapis.com/auth/bigquery"`. We should only need 1, and per the rest of cloud we should use `"https://www.googleapis.com/auth/cloud-platform"` Note: I'm filing this issue due to https://github.com/googleapis/google-auth-library-python/issues/1204. I believe that even after the issue identified there is addressed (default scopes shouldn't cause user/impersonated credentials to downscope), google-auth might be downscoping credentials unnecessarily because it sees the "bigquery" scope and it's not present by default.
1.0
"https://www.googleapis.com/auth/bigquery" scope is unecessary - https://github.com/googleapis/python-bigquery/blob/e1aa9218ad22f85c9a6cab8b61d013779376a582/google/cloud/bigquery/client.py#L229 `"https://www.googleapis.com/auth/cloud-platform"` is a superset of `"https://www.googleapis.com/auth/bigquery"`. We should only need 1, and per the rest of cloud we should use `"https://www.googleapis.com/auth/cloud-platform"` Note: I'm filing this issue due to https://github.com/googleapis/google-auth-library-python/issues/1204. I believe that even after the issue identified there is addressed (default scopes shouldn't cause user/impersonated credentials to downscope), google-auth might be downscoping credentials unnecessarily because it sees the "bigquery" scope and it's not present by default.
process
scope is unecessary is a superset of we should only need and per the rest of cloud we should use note i m filing this issue due to i believe that even after the issue identified there is addressed default scopes shouldn t cause user impersonated credentials to downscope google auth might be downscoping credentials unnecessarily because it sees the bigquery scope and it s not present by default
1
6,111
8,969,727,395
IssuesEvent
2019-01-29 11:38:28
ec-europa/europa-component-library
https://api.github.com/repos/ec-europa/europa-component-library
opened
[RFC] File component - INNO-1320
Improvement RFC process: WIP
More attributes can be used on the download links in the file component: https://v2--europa-component-library.netlify.com/ec/components/file/code/ - `download` https://www.w3schools.com/tags/att_a_download.asp - `hreflang` https://www.w3schools.com/tags/att_a_hreflang.asp - `lang` https://www.w3schools.com/tags/att_global_lang.asp which is not specific to links but should be used here since the label of the link is in another language
1.0
[RFC] File component - INNO-1320 - More attributes can be used on the download links in the file component: https://v2--europa-component-library.netlify.com/ec/components/file/code/ - `download` https://www.w3schools.com/tags/att_a_download.asp - `hreflang` https://www.w3schools.com/tags/att_a_hreflang.asp - `lang` https://www.w3schools.com/tags/att_global_lang.asp which is not specific to links but should be used here since the label of the link is in another language
process
file component inno more attributes can be used on the download links in the file component download hreflang lang which is not specific to links but should be used here since the label of the link is in another language
1
12,453
14,935,214,128
IssuesEvent
2021-01-25 11:37:16
smertatli/SWE-573
https://api.github.com/repos/smertatli/SWE-573
closed
Requirement Elicitation
done in process
Meet with the clients to elicit the requirements for the project to be done. - What is the business problem we will solve by building this software product? - Who are the people that need this solution? - Who are the key stakeholders? - Why does such a problem exist? - What does this product improve? - How would you quantify success of this product? - What is the most important output of this product for you?
1.0
Requirement Elicitation - Meet with the clients to elicit the requirements for the project to be done. - What is the business problem we will solve by building this software product? - Who are the people that need this solution? - Who are the key stakeholders? - Why does such a problem exist? - What does this product improve? - How would you quantify success of this product? - What is the most important output of this product for you?
process
requirement elicitation meet with the clients to elicit the requirements for the project to be done what is the business problem we will solve by building this software product who are the people that need this solution who are the key stakeholders why does such a problem exist what does this product improve how would you quantify success of this product what is the most important output of this product for you
1
75,681
3,470,980,354
IssuesEvent
2015-12-23 12:25:08
USGS-WiM/SiGLDMS
https://api.github.com/repos/USGS-WiM/SiGLDMS
closed
Organization "Add new" button (CSS)
Priority: Medium
Changed from "Add New" string to "Add New" button. Needs styling. ![neworgbutton](https://cloud.githubusercontent.com/assets/1580076/11509567/6185359a-9824-11e5-9072-26331520377c.JPG)
1.0
Organization "Add new" button (CSS) - Changed from "Add New" string to "Add New" button. Needs styling. ![neworgbutton](https://cloud.githubusercontent.com/assets/1580076/11509567/6185359a-9824-11e5-9072-26331520377c.JPG)
non_process
organization add new button css changed from add new string to add new button needs styling
0
142,019
19,012,457,231
IssuesEvent
2021-11-23 10:48:02
Yann-dv/_Pekocko
https://api.github.com/repos/Yann-dv/_Pekocko
opened
CVE-2020-7608 (Medium) detected in yargs-parser-7.0.0.tgz, yargs-parser-10.1.0.tgz
security vulnerability
## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-7.0.0.tgz</b>, <b>yargs-parser-10.1.0.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p> <p>Path to dependency file: _Pekocko/package.json</p> <p>Path to vulnerable library: _Pekocko/node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - compiler-cli-7.0.0.tgz (Root Library) - yargs-9.0.1.tgz - :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p> <p>Path to dependency file: _Pekocko/package.json</p> <p>Path to vulnerable library: _Pekocko/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.10.2.tgz (Root Library) - webpack-dev-server-3.1.8.tgz - yargs-12.0.2.tgz - :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Yann-dv/_Pekocko/commit/29a980e4dad903d391a0354b9cb7c71642e2c2fe">29a980e4dad903d391a0354b9cb7c71642e2c2fe</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7608 (Medium) detected in yargs-parser-7.0.0.tgz, yargs-parser-10.1.0.tgz - ## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-7.0.0.tgz</b>, <b>yargs-parser-10.1.0.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p> <p>Path to dependency file: _Pekocko/package.json</p> <p>Path to vulnerable library: _Pekocko/node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - compiler-cli-7.0.0.tgz (Root Library) - yargs-9.0.1.tgz - :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p> <p>Path to dependency file: _Pekocko/package.json</p> <p>Path to vulnerable library: _Pekocko/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.10.2.tgz (Root Library) - webpack-dev-server-3.1.8.tgz - yargs-12.0.2.tgz - :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Yann-dv/_Pekocko/commit/29a980e4dad903d391a0354b9cb7c71642e2c2fe">29a980e4dad903d391a0354b9cb7c71642e2c2fe</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in yargs parser tgz yargs parser tgz cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file pekocko package json path to vulnerable library pekocko node modules angular compiler cli node modules yargs parser package json dependency hierarchy compiler cli tgz root library yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file pekocko package json path to vulnerable library pekocko node modules webpack dev server node modules yargs parser package json dependency hierarchy build angular tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch main vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
19,337
4,383,043,706
IssuesEvent
2016-08-07 08:58:24
zcash/zcash
https://api.github.com/repos/zcash/zcash
closed
Finish the list of consensus changes for security auditors
consensus protocol documentation SECURITY
The security auditors need the [list of consensus changes](https://github.com/zcash/zcash/wiki/Security-Auditor-Quick-Start#list-of-consensus-changes) to be finished before the audit can begin. It would also be nice if those bullet points linked to relevant code (in the security-review-frozen version).
1.0
Finish the list of consensus changes for security auditors - The security auditors need the [list of consensus changes](https://github.com/zcash/zcash/wiki/Security-Auditor-Quick-Start#list-of-consensus-changes) to be finished before the audit can begin. It would also be nice if those bullet points linked to relevant code (in the security-review-frozen version).
non_process
finish the list of consensus changes for security auditors the security auditors need the to be finished before the audit can begin it would also be nice if those bullet points linked to relevant code in the security review frozen version
0
2,645
5,425,336,484
IssuesEvent
2017-03-03 05:37:46
FujiXeroxNZ-Wellington/Indigo
https://api.github.com/repos/FujiXeroxNZ-Wellington/Indigo
closed
js Auto validate cannot validate elements in List group
0-4-Contract Processing 0-Contract Management bug v1.0
Error: Angular-auto-validate: invalid bs3 form structure elements must be wrapped by a form-group class the above error is thrown when validating the form fields using js-autovalidate library
1.0
js Auto validate cannot validate elements in List group - Error: Angular-auto-validate: invalid bs3 form structure elements must be wrapped by a form-group class the above error is thrown when validating the form fields using js-autovalidate library
process
js auto validate cannot validate elements in list group error angular auto validate invalid form structure elements must be wrapped by a form group class the above error is thrown when validating the form fields using js autovalidate library
1
156,230
19,831,499,138
IssuesEvent
2022-01-20 12:29:48
Dima2021/SecurityShepherd
https://api.github.com/repos/Dima2021/SecurityShepherd
closed
CVE-2020-9493 (High) detected in log4j-1.2.7.jar
security vulnerability
## CVE-2020-9493 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.7.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /tory/log4j/log4j/1.2.7/log4j-1.2.7.jar,/target/owaspSecurityShepherd/WEB-INF/lib/log4j-1.2.7.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.7.jar** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution. <p>Publish Date: 2021-06-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493>CVE-2020-9493</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.7","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.7","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-9493","vulnerabilityDetails":"A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-9493 (High) detected in log4j-1.2.7.jar - ## CVE-2020-9493 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.7.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /tory/log4j/log4j/1.2.7/log4j-1.2.7.jar,/target/owaspSecurityShepherd/WEB-INF/lib/log4j-1.2.7.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.7.jar** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution. <p>Publish Date: 2021-06-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493>CVE-2020-9493</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.7","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.7","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-9493","vulnerabilityDetails":"A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9493","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jar cve high severity vulnerability vulnerable library jar path to dependency file pom xml path to vulnerable library tory jar target owaspsecurityshepherd web inf lib jar dependency hierarchy x jar vulnerable library found in base branch dev vulnerability details a deserialization flaw was found in apache chainsaw versions prior to which could lead to malicious code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a deserialization flaw was found in apache chainsaw versions prior to which could lead to malicious code execution vulnerabilityurl
0
11,510
14,394,750,249
IssuesEvent
2020-12-03 02:02:31
A01551343/4a
https://api.github.com/repos/A01551343/4a
closed
complete_size_estimating_template
process-dashboard
- completar el formato de estimacion de LOC con los valores reales obtenidos
1.0
complete_size_estimating_template - - completar el formato de estimacion de LOC con los valores reales obtenidos
process
complete size estimating template completar el formato de estimacion de loc con los valores reales obtenidos
1
140,628
18,905,973,963
IssuesEvent
2021-11-16 09:08:01
VerdantSparks/vuetify_ts_aspnetcore_starter
https://api.github.com/repos/VerdantSparks/vuetify_ts_aspnetcore_starter
closed
CVE-2018-8292 (High) detected in microsoft.netcore.app.2.0.0.nupkg, system.net.http.4.3.0.nupkg
security vulnerability
## CVE-2018-8292 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>microsoft.netcore.app.2.0.0.nupkg</b>, <b>system.net.http.4.3.0.nupkg</b></p></summary> <p> <details><summary><b>microsoft.netcore.app.2.0.0.nupkg</b></p></summary> <p>A set of .NET API's that are included in the default .NET Core application model. e8b8861ac7faf042c...</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg</a></p> <p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p> <p>Path to vulnerable library: ft.netcore.app/2.0.0/microsoft.netcore.app.2.0.0.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.netcore.app.2.0.0.nupkg** (Vulnerable Library) </details> <details><summary><b>system.net.http.4.3.0.nupkg</b></p></summary> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p> <p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - microsoft.aspnetcore.authentication.jwtbearer.2.0.4.nupkg (Root Library) - microsoft.identitymodel.protocols.openidconnect.2.1.4.nupkg - microsoft.identitymodel.protocols.2.1.4.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/VerdantSparks/vuetify_ts_aspnetcore_starter/commit/898883a08965de37520ac5536d5fba3acbd14129">898883a08965de37520ac5536d5fba3acbd14129</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0. <p>Publish Date: 2018-10-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8292>CVE-2018-8292</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/dotnet/announcements/issues/88">https://github.com/dotnet/announcements/issues/88</a></p> <p>Release Date: 2018-10-10</p> <p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-8292 (High) detected in microsoft.netcore.app.2.0.0.nupkg, system.net.http.4.3.0.nupkg - ## CVE-2018-8292 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>microsoft.netcore.app.2.0.0.nupkg</b>, <b>system.net.http.4.3.0.nupkg</b></p></summary> <p> <details><summary><b>microsoft.netcore.app.2.0.0.nupkg</b></p></summary> <p>A set of .NET API's that are included in the default .NET Core application model. e8b8861ac7faf042c...</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg">https://api.nuget.org/packages/microsoft.netcore.app.2.0.0.nupkg</a></p> <p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p> <p>Path to vulnerable library: ft.netcore.app/2.0.0/microsoft.netcore.app.2.0.0.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.netcore.app.2.0.0.nupkg** (Vulnerable Library) </details> <details><summary><b>system.net.http.4.3.0.nupkg</b></p></summary> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: vuetify_ts_aspnetcore_starter/vuetify_ts_aspnetcore_starter.csproj</p> <p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - microsoft.aspnetcore.authentication.jwtbearer.2.0.4.nupkg (Root Library) - microsoft.identitymodel.protocols.openidconnect.2.1.4.nupkg - microsoft.identitymodel.protocols.2.1.4.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/VerdantSparks/vuetify_ts_aspnetcore_starter/commit/898883a08965de37520ac5536d5fba3acbd14129">898883a08965de37520ac5536d5fba3acbd14129</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0. <p>Publish Date: 2018-10-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8292>CVE-2018-8292</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/dotnet/announcements/issues/88">https://github.com/dotnet/announcements/issues/88</a></p> <p>Release Date: 2018-10-10</p> <p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in microsoft netcore app nupkg system net http nupkg cve high severity vulnerability vulnerable libraries microsoft netcore app nupkg system net http nupkg microsoft netcore app nupkg a set of net api s that are included in the default net core application model library home page a href path to dependency file vuetify ts aspnetcore starter vuetify ts aspnetcore starter csproj path to vulnerable library ft netcore app microsoft netcore app nupkg dependency hierarchy x microsoft netcore app nupkg vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file vuetify ts aspnetcore starter vuetify ts aspnetcore starter csproj path to vulnerable library usr share dotnet sdk nugetfallbackfolder system net http system net http nupkg dependency hierarchy microsoft aspnetcore authentication jwtbearer nupkg root library microsoft identitymodel protocols openidconnect nupkg microsoft identitymodel protocols nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch master vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system net http microsoft powershell commands utility rc step up your open source security game with whitesource
0
67,236
8,114,933,365
IssuesEvent
2018-08-15 03:36:11
the-tale/the-tale
https://api.github.com/repos/the-tale/the-tale
closed
Карты судьбы: поправить карты прямого начисления влияния
comp_cards comp_politics cont_game_designe est_simple type_improvement
Дополнить описание перечнем модификаторов, которые должны влиять на начисляемое влияние. Дописать тесты, которые будут это проверять. Должны ли хоть какие-то модификаторы влиять на эту величину? В текущий момент получается, что есть карты на влияние в задании, на которые действуют все бонусы и штрафы. А есть карты прямого влияния, которые, по идее, должны просто гарантированно давать указанную сумму. Возможно, надо изменить количество начисляемого влияняи, чтобы компенсировать отсутствие бонусов.
1.0
Карты судьбы: поправить карты прямого начисления влияния - Дополнить описание перечнем модификаторов, которые должны влиять на начисляемое влияние. Дописать тесты, которые будут это проверять. Должны ли хоть какие-то модификаторы влиять на эту величину? В текущий момент получается, что есть карты на влияние в задании, на которые действуют все бонусы и штрафы. А есть карты прямого влияния, которые, по идее, должны просто гарантированно давать указанную сумму. Возможно, надо изменить количество начисляемого влияняи, чтобы компенсировать отсутствие бонусов.
non_process
карты судьбы поправить карты прямого начисления влияния дополнить описание перечнем модификаторов которые должны влиять на начисляемое влияние дописать тесты которые будут это проверять должны ли хоть какие то модификаторы влиять на эту величину в текущий момент получается что есть карты на влияние в задании на которые действуют все бонусы и штрафы а есть карты прямого влияния которые по идее должны просто гарантированно давать указанную сумму возможно надо изменить количество начисляемого влияняи чтобы компенсировать отсутствие бонусов
0
135,257
30,274,098,750
IssuesEvent
2023-07-07 17:55:12
WordPress/openverse
https://api.github.com/repos/WordPress/openverse
closed
Use Elasticsearch 8 locally
🟧 priority: high 💻 aspect: code 🧰 goal: internal improvement 🧱 stack: api 🔧 tech: elasticsearch
## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> We need to make sure Openverse will work with ElasticSearch 8 prior to upgrading our existing cluster. ## Description <!-- Describe the feature and how it solves the problem. --> Update the root docker compose to use version `8.8.2` of ElasticSearch, or whatever the latest version is at the time of experimentation. https://github.com/WordPress/openverse/blob/2f39baa19fecbbf26fe3be913770c26fe9d57794/docker-compose.yml#L184 Additionally we should carefully review the [migration guide](https://www.elastic.co/guide/en/elasticsearch/reference/current/migrating-8.0.html) to make sure nothing we use has been deprecated or altered. ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> <!-- If you would like to work on this, please comment below separately. -->
1.0
Use Elasticsearch 8 locally - ## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> We need to make sure Openverse will work with ElasticSearch 8 prior to upgrading our existing cluster. ## Description <!-- Describe the feature and how it solves the problem. --> Update the root docker compose to use version `8.8.2` of ElasticSearch, or whatever the latest version is at the time of experimentation. https://github.com/WordPress/openverse/blob/2f39baa19fecbbf26fe3be913770c26fe9d57794/docker-compose.yml#L184 Additionally we should carefully review the [migration guide](https://www.elastic.co/guide/en/elasticsearch/reference/current/migrating-8.0.html) to make sure nothing we use has been deprecated or altered. ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> <!-- If you would like to work on this, please comment below separately. -->
non_process
use elasticsearch locally problem we need to make sure openverse will work with elasticsearch prior to upgrading our existing cluster description update the root docker compose to use version of elasticsearch or whatever the latest version is at the time of experimentation additionally we should carefully review the to make sure nothing we use has been deprecated or altered alternatives additional context
0
376,908
26,222,514,223
IssuesEvent
2023-01-04 15:53:27
SandraScherer/EntertainmentInfothek
https://api.github.com/repos/SandraScherer/EntertainmentInfothek
closed
Add sound mix information to series
documentation enhancement database program
- [x] Add table Series_SoundMix to database - [x] Add/adapt Series class in EntertainmentDB.dll - [x] Add tests to EntertainmentDB.Tests - [x] Add/adapt ContentCreator classes in WikiPageCreator - [x] Add tests to WikiPageCreator.Tests - [x] Update documentation - [x] EntertainmentInfothek_Database.vpp - [x] EntertainmentInfothek_EntertainmentDB.dll.vpp - [x] EntertainmentInfothek_WikiPageCreator.vpp - [x] Doxygen
1.0
Add sound mix information to series - - [x] Add table Series_SoundMix to database - [x] Add/adapt Series class in EntertainmentDB.dll - [x] Add tests to EntertainmentDB.Tests - [x] Add/adapt ContentCreator classes in WikiPageCreator - [x] Add tests to WikiPageCreator.Tests - [x] Update documentation - [x] EntertainmentInfothek_Database.vpp - [x] EntertainmentInfothek_EntertainmentDB.dll.vpp - [x] EntertainmentInfothek_WikiPageCreator.vpp - [x] Doxygen
non_process
add sound mix information to series add table series soundmix to database add adapt series class in entertainmentdb dll add tests to entertainmentdb tests add adapt contentcreator classes in wikipagecreator add tests to wikipagecreator tests update documentation entertainmentinfothek database vpp entertainmentinfothek entertainmentdb dll vpp entertainmentinfothek wikipagecreator vpp doxygen
0
15,408
19,598,022,471
IssuesEvent
2022-01-05 20:27:44
opensearch-project/data-prepper
https://api.github.com/repos/opensearch-project/data-prepper
opened
Create an Aggregate Processor
plugin - processor
**Is your feature request related to a problem? Please describe.** Data Prepper users need a processor that can be utilized for generic stateful aggregation **Describe the solution you'd like** A solution has already been proposed in RFC #699
1.0
Create an Aggregate Processor - **Is your feature request related to a problem? Please describe.** Data Prepper users need a processor that can be utilized for generic stateful aggregation **Describe the solution you'd like** A solution has already been proposed in RFC #699
process
create an aggregate processor is your feature request related to a problem please describe data prepper users need a processor that can be utilized for generic stateful aggregation describe the solution you d like a solution has already been proposed in rfc
1
46,008
18,924,532,155
IssuesEvent
2021-11-17 08:02:01
Azure/azure-powershell
https://api.github.com/repos/Azure/azure-powershell
closed
Remove-AzSqlDatabaseSecondary Throws Object reference not set to an instance of an object
SQL Service Attention question customer-reported needs-author-feedback no-recent-activity
<!-- - Make sure you are able to reproduce this issue on the latest released version of Az - https://www.powershellgallery.com/packages/Az - Please search the existing issues to see if there has been a similar issue filed - For issue related to importing a module, please refer to our troubleshooting guide: - https://github.com/Azure/azure-powershell/blob/master/documentation/troubleshoot-module-load.md --> ## Description Cmdlet fails with both parameters and literals. ## Steps to reproduce ```powershell Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName $svr.ResourceGroupName -PartnerServerName $svr.ServerName -DatabaseName $drdbname -servername $targetSvr -ResourceGroupName $targetRG ``` ## Environment data <!-- Please run $PSVersionTable and paste the output in the below code block If running the Docker container image, indicate the tag of the image used and the version of Docker engine--> ``` ``` ## Module versions <!-- Please run (Get-Module -ListAvailable) and paste the output in the below code block --> ```powershell This is failing the same way on both Az.SQL 2.12.0 and 2.16.0 PS /Users/joey> $PsVersionTable Name Value ---- ----- PSVersion 7.1.2 PSEdition Core GitCommitId 7.1.2 OS Darwin 20.3.0 Darwin Kernel Version 20.3.0: Thu Jan 21 00:07:06 PST 2021; root:xnu-7195.81.3~1/RELEASE_X86_64 Platform Unix PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…} PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 WSManStackVersion 3.0 and cloud shell ``` ## Debug output <!-- Set $DebugPreference='Continue' before running the repro and paste the resulting debug stream in the below code block ATTENTION: Be sure to remove any sensitive information that may be in the logs --> ``` Message : Object reference not set to an instance of an object. StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord() Exception : System.NullReferenceException InvocationInfo : {Remove-AzSqlDatabaseSecondary} Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' -PartnerServerName 'dcac-demo' -DatabaseName 'adventureworks2017-dr' -servername 'dcac-demo-dr' -ResourceGroupName 'dcac-sqldr' -verbose -debug Position : At line:1 char:1 + Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' … + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ HistoryId : 33 HistoryId: 32 Message : Object reference not set to an instance of an object. StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord() Exception : System.NullReferenceException InvocationInfo : {Remove-AzSqlDatabaseSecondary} Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug Position : At line:1 char:1 + Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - … + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ HistoryId : 32 HistoryId: 30 Message : Object reference not set to an instance of an object. StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord() Exception : System.NullReferenceException InvocationInfo : {Remove-AzSqlDatabaseSecondary} Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug Position : At line:1 char:1 + Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - … + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ HistoryId : 30 ``` ## Error output <!-- Please run Resolve-AzError and paste the output in the below code block ATTENTION: Be sure to remove any sensitive information that may be in the logs --> ``` Remove-AzSqlDatabaseSecondary: Object reference not set to an instance of an object. DEBUG: AzureQoSEvent: CommandName - Remove-AzSqlDatabaseSecondary; IsSuccess - False; Duration - 00:00:00.8716089;; Exception - System.NullReferenceException: Object reference not set to an instance of an object. at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord(); DEBUG: Finish sending metric. DEBUG: 4:43:03 PM - RemoveAzureSqlDatabaseSecondary end processing. ```
1.0
Remove-AzSqlDatabaseSecondary Throws Object reference not set to an instance of an object - <!-- - Make sure you are able to reproduce this issue on the latest released version of Az - https://www.powershellgallery.com/packages/Az - Please search the existing issues to see if there has been a similar issue filed - For issue related to importing a module, please refer to our troubleshooting guide: - https://github.com/Azure/azure-powershell/blob/master/documentation/troubleshoot-module-load.md --> ## Description Cmdlet fails with both parameters and literals. ## Steps to reproduce ```powershell Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName $svr.ResourceGroupName -PartnerServerName $svr.ServerName -DatabaseName $drdbname -servername $targetSvr -ResourceGroupName $targetRG ``` ## Environment data <!-- Please run $PSVersionTable and paste the output in the below code block If running the Docker container image, indicate the tag of the image used and the version of Docker engine--> ``` ``` ## Module versions <!-- Please run (Get-Module -ListAvailable) and paste the output in the below code block --> ```powershell This is failing the same way on both Az.SQL 2.12.0 and 2.16.0 PS /Users/joey> $PsVersionTable Name Value ---- ----- PSVersion 7.1.2 PSEdition Core GitCommitId 7.1.2 OS Darwin 20.3.0 Darwin Kernel Version 20.3.0: Thu Jan 21 00:07:06 PST 2021; root:xnu-7195.81.3~1/RELEASE_X86_64 Platform Unix PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…} PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 WSManStackVersion 3.0 and cloud shell ``` ## Debug output <!-- Set $DebugPreference='Continue' before running the repro and paste the resulting debug stream in the below code block ATTENTION: Be sure to remove any sensitive information that may be in the logs --> ``` Message : Object reference not set to an instance of an object. StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord() Exception : System.NullReferenceException InvocationInfo : {Remove-AzSqlDatabaseSecondary} Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' -PartnerServerName 'dcac-demo' -DatabaseName 'adventureworks2017-dr' -servername 'dcac-demo-dr' -ResourceGroupName 'dcac-sqldr' -verbose -debug Position : At line:1 char:1 + Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName 'sqldb-demo' … + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ HistoryId : 33 HistoryId: 32 Message : Object reference not set to an instance of an object. StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord() Exception : System.NullReferenceException InvocationInfo : {Remove-AzSqlDatabaseSecondary} Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug Position : At line:1 char:1 + Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - … + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ HistoryId : 32 HistoryId: 30 Message : Object reference not set to an instance of an object. StackTrace : at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord() Exception : System.NullReferenceException InvocationInfo : {Remove-AzSqlDatabaseSecondary} Line : Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo -PartnerServerName dcac-demo -DatabaseName adventureworks2017-dr -servername dcac-demo-dr -ResourceGroupName dcac-sqldr -verbose -debug Position : At line:1 char:1 + Remove-AzSqlDatabaseSecondary -PartnerResourceGroupName sqldb-demo - … + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ HistoryId : 30 ``` ## Error output <!-- Please run Resolve-AzError and paste the output in the below code block ATTENTION: Be sure to remove any sensitive information that may be in the logs --> ``` Remove-AzSqlDatabaseSecondary: Object reference not set to an instance of an object. DEBUG: AzureQoSEvent: CommandName - Remove-AzSqlDatabaseSecondary; IsSuccess - False; Duration - 00:00:00.8716089;; Exception - System.NullReferenceException: Object reference not set to an instance of an object. at Microsoft.Azure.Commands.Sql.ReplicationLink.Services.AzureSqlDatabaseReplicationAdapter.RemoveLink(String resourceGroupName, String serverName, String databaseName, String partnerResourceGroupName, String partnerServerName) at Microsoft.Azure.Commands.Sql.Replication.Cmdlet.RemoveAzureSqlDatabaseSecondary.PersistChanges(IEnumerable`1 entity) at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.<>c__DisplayClass16_0.<ExecuteCmdlet>b__0() at Microsoft.Azure.Commands.Sql.Common.AzureSqlCmdletBase`2.ExecuteCmdlet() at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord(); DEBUG: Finish sending metric. DEBUG: 4:43:03 PM - RemoveAzureSqlDatabaseSecondary end processing. ```
non_process
remove azsqldatabasesecondary throws object reference not set to an instance of an object make sure you are able to reproduce this issue on the latest released version of az please search the existing issues to see if there has been a similar issue filed for issue related to importing a module please refer to our troubleshooting guide description cmdlet fails with both parameters and literals steps to reproduce powershell remove azsqldatabasesecondary partnerresourcegroupname svr resourcegroupname partnerservername svr servername databasename drdbname servername targetsvr resourcegroupname targetrg environment data please run psversiontable and paste the output in the below code block if running the docker container image indicate the tag of the image used and the version of docker engine module versions powershell this is failing the same way on both az sql and ps users joey psversiontable name value psversion psedition core gitcommitid os darwin darwin kernel version thu jan pst root xnu release platform unix pscompatibleversions … psremotingprotocolversion serializationversion wsmanstackversion and cloud shell debug output set debugpreference continue before running the repro and paste the resulting debug stream in the below code block attention be sure to remove any sensitive information that may be in the logs message object reference not set to an instance of an object stacktrace at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception system nullreferenceexception invocationinfo remove azsqldatabasesecondary line remove azsqldatabasesecondary partnerresourcegroupname sqldb demo partnerservername dcac demo databasename dr servername dcac demo dr resourcegroupname dcac sqldr verbose debug position at line char remove azsqldatabasesecondary partnerresourcegroupname sqldb demo … historyid historyid message object reference not set to an instance of an object stacktrace at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft windowsazure commands utilities common azurepscmdlet confirmaction string processmessage string target action action at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception system nullreferenceexception invocationinfo remove azsqldatabasesecondary line remove azsqldatabasesecondary partnerresourcegroupname sqldb demo partnerservername dcac demo databasename dr servername dcac demo dr resourcegroupname dcac sqldr verbose debug position at line char remove azsqldatabasesecondary partnerresourcegroupname sqldb demo … historyid historyid message object reference not set to an instance of an object stacktrace at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft windowsazure commands utilities common azurepscmdlet confirmaction string processmessage string target action action at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception system nullreferenceexception invocationinfo remove azsqldatabasesecondary line remove azsqldatabasesecondary partnerresourcegroupname sqldb demo partnerservername dcac demo databasename dr servername dcac demo dr resourcegroupname dcac sqldr verbose debug position at line char remove azsqldatabasesecondary partnerresourcegroupname sqldb demo … historyid error output please run resolve azerror and paste the output in the below code block attention be sure to remove any sensitive information that may be in the logs remove azsqldatabasesecondary object reference not set to an instance of an object debug azureqosevent commandname remove azsqldatabasesecondary issuccess false duration exception system nullreferenceexception object reference not set to an instance of an object at microsoft azure commands sql replicationlink services azuresqldatabasereplicationadapter removelink string resourcegroupname string servername string databasename string partnerresourcegroupname string partnerservername at microsoft azure commands sql replication cmdlet removeazuresqldatabasesecondary persistchanges ienumerable entity at microsoft azure commands sql common azuresqlcmdletbase c b at microsoft azure commands sql common azuresqlcmdletbase executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord debug finish sending metric debug pm removeazuresqldatabasesecondary end processing
0
9,923
12,963,288,340
IssuesEvent
2020-07-20 18:32:26
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Passing parameters between triggering pipelines
Pri2 devops-cicd-process/tech devops/prod support-request
How do I pass (or receive) parameters between pipelines? My triggered pipeline has input parameters that need to be set from the 'build' pipeline that triggers it. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 86285f72-9e28-da97-59bb-c29eb60f627d * Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54 * Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @ashkir * Microsoft Alias: **ashkir**
1.0
Passing parameters between triggering pipelines - How do I pass (or receive) parameters between pipelines? My triggered pipeline has input parameters that need to be set from the 'build' pipeline that triggers it. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 86285f72-9e28-da97-59bb-c29eb60f627d * Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54 * Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @ashkir * Microsoft Alias: **ashkir**
process
passing parameters between triggering pipelines how do i pass or receive parameters between pipelines my triggered pipeline has input parameters that need to be set from the build pipeline that triggers it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login ashkir microsoft alias ashkir
1
4,410
7,299,077,731
IssuesEvent
2018-02-26 18:58:08
UKHomeOffice/dq-aws-transition
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
closed
Finalise Windows Data Ingest Prod server
DQ Data Ingest Production S4 Processing
Finalise Windows Data Ingest Prod server - [x] Terraform Deployment - [x] Post-Deployment Tasks - [x] SEQ check running - [x] Data transfer tasks running - [x] Swapped in vars ## Acceptance criteria - [x] Data is flowing in from S3 - [x] Data is being processed by SEQ job - [x] WS is launching jobs - [x] Data is flowing out to S3
1.0
Finalise Windows Data Ingest Prod server - Finalise Windows Data Ingest Prod server - [x] Terraform Deployment - [x] Post-Deployment Tasks - [x] SEQ check running - [x] Data transfer tasks running - [x] Swapped in vars ## Acceptance criteria - [x] Data is flowing in from S3 - [x] Data is being processed by SEQ job - [x] WS is launching jobs - [x] Data is flowing out to S3
process
finalise windows data ingest prod server finalise windows data ingest prod server terraform deployment post deployment tasks seq check running data transfer tasks running swapped in vars acceptance criteria data is flowing in from data is being processed by seq job ws is launching jobs data is flowing out to
1
418,698
12,202,203,356
IssuesEvent
2020-04-30 08:34:35
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 staging-1528] Work party: strange template for the first party
Priority: Medium Status: Fixed
1. Make a new server 2. Place a workbench and make 10 hewn log order 3. Place a contract board and start to create workparty You will see this strange template ![bandicam 2020-04-28 14-24-17-858](https://user-images.githubusercontent.com/27898520/80470998-3789dd80-895c-11ea-87e9-c1ed6fb87f25.jpg)
1.0
[0.9.0 staging-1528] Work party: strange template for the first party - 1. Make a new server 2. Place a workbench and make 10 hewn log order 3. Place a contract board and start to create workparty You will see this strange template ![bandicam 2020-04-28 14-24-17-858](https://user-images.githubusercontent.com/27898520/80470998-3789dd80-895c-11ea-87e9-c1ed6fb87f25.jpg)
non_process
work party strange template for the first party make a new server place a workbench and make hewn log order place a contract board and start to create workparty you will see this strange template
0
55,615
13,647,458,165
IssuesEvent
2020-09-26 03:25:41
TerryCavanagh/diceydungeonsbeta
https://api.github.com/repos/TerryCavanagh/diceydungeonsbeta
closed
UI Overlap When Gold Appears in Combat
High Priority v0.5: 21st June Build
This was fixed [here](https://github.com/TerryCavanagh/diceydungeons.com/issues/755) but is back again in Witch PU. ![overlap](https://user-images.githubusercontent.com/40739964/60115151-481d6600-9743-11e9-8b4a-b4c10d6c7ef0.jpg)
1.0
UI Overlap When Gold Appears in Combat - This was fixed [here](https://github.com/TerryCavanagh/diceydungeons.com/issues/755) but is back again in Witch PU. ![overlap](https://user-images.githubusercontent.com/40739964/60115151-481d6600-9743-11e9-8b4a-b4c10d6c7ef0.jpg)
non_process
ui overlap when gold appears in combat this was fixed but is back again in witch pu
0
32,969
13,994,311,777
IssuesEvent
2020-10-28 00:20:57
Azure/azure-sdk-for-java
https://api.github.com/repos/Azure/azure-sdk-for-java
closed
[Service Bus] Fix module-info.java
Client Service Bus blocking-release
`module-info.java` in Service Bus has many `opens` statements exposing multiple packages. These `opens` statements have to be removed before GA.
1.0
[Service Bus] Fix module-info.java - `module-info.java` in Service Bus has many `opens` statements exposing multiple packages. These `opens` statements have to be removed before GA.
non_process
fix module info java module info java in service bus has many opens statements exposing multiple packages these opens statements have to be removed before ga
0
43,343
11,628,285,965
IssuesEvent
2020-02-27 18:02:10
google/flogger
https://api.github.com/repos/google/flogger
closed
Flag --incompatible_load_java_rules_from_bzl will break Flogger in Bazel 1.2.1
P3 type=defect
Incompatible flag --incompatible_load_java_rules_from_bzl will break Flogger once Bazel 1.2.1 is released. Please see the following CI builds for more information: * [:ubuntu: 16.04 (OpenJDK 8)](<a href="https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/342#7fdfdbb2-c0ac-49f7-8fb4-4af9267c90e6" target="_blank">:ubuntu: 16.04 (OpenJDK 8)</a>) Questions? Please file an issue in https://github.com/bazelbuild/continuous-integration **Important**: Please do NOT modify the issue title since that might break our tools.
1.0
Flag --incompatible_load_java_rules_from_bzl will break Flogger in Bazel 1.2.1 - Incompatible flag --incompatible_load_java_rules_from_bzl will break Flogger once Bazel 1.2.1 is released. Please see the following CI builds for more information: * [:ubuntu: 16.04 (OpenJDK 8)](<a href="https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/342#7fdfdbb2-c0ac-49f7-8fb4-4af9267c90e6" target="_blank">:ubuntu: 16.04 (OpenJDK 8)</a>) Questions? Please file an issue in https://github.com/bazelbuild/continuous-integration **Important**: Please do NOT modify the issue title since that might break our tools.
non_process
flag incompatible load java rules from bzl will break flogger in bazel incompatible flag incompatible load java rules from bzl will break flogger once bazel is released please see the following ci builds for more information ubuntu openjdk questions please file an issue in important please do not modify the issue title since that might break our tools
0
7,619
10,727,398,402
IssuesEvent
2019-10-28 11:36:08
prisma/photonjs
https://api.github.com/repos/prisma/photonjs
closed
Filtering `null` relationships is not possible
bug/2-confirmed kind/bug process/candidate
## Schema ```prisma model User { id Int @id address Address? } model Address { id Int @id } ``` ## Code ```ts const usersWithoutAddress = await photon.users.findMany({ where: { address: null } }) ``` This generates an incorrect query under the hood with querying `address: {}` instead of `address: null`.
1.0
Filtering `null` relationships is not possible - ## Schema ```prisma model User { id Int @id address Address? } model Address { id Int @id } ``` ## Code ```ts const usersWithoutAddress = await photon.users.findMany({ where: { address: null } }) ``` This generates an incorrect query under the hood with querying `address: {}` instead of `address: null`.
process
filtering null relationships is not possible schema prisma model user id int id address address model address id int id code ts const userswithoutaddress await photon users findmany where address null this generates an incorrect query under the hood with querying address instead of address null
1
17,560
23,372,954,891
IssuesEvent
2022-08-10 21:54:33
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Tip: You can create an empty environment and reference it from deployment jobs
devops/prod cba Pri1 devops-cicd-process/tech
The document contains this tip: "Tip: You can create an empty environment and reference it from deployment jobs" I would like to do that, meaning I would like to run a deployment job that does not actually move any files but just records the deployment. However, when I try, I get the error: "No resource were found in the environment with ID 24 matching the specified criteria: ResourceId , ResourceName , ResourceType VirtualMachine, Tags ." Where can I find more information about using empty environments? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Create target environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Tip: You can create an empty environment and reference it from deployment jobs - The document contains this tip: "Tip: You can create an empty environment and reference it from deployment jobs" I would like to do that, meaning I would like to run a deployment job that does not actually move any files but just records the deployment. However, when I try, I get the error: "No resource were found in the environment with ID 24 matching the specified criteria: ResourceId , ResourceName , ResourceType VirtualMachine, Tags ." Where can I find more information about using empty environments? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Create target environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
tip you can create an empty environment and reference it from deployment jobs the document contains this tip tip you can create an empty environment and reference it from deployment jobs i would like to do that meaning i would like to run a deployment job that does not actually move any files but just records the deployment however when i try i get the error no resource were found in the environment with id matching the specified criteria resourceid resourcename resourcetype virtualmachine tags where can i find more information about using empty environments document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
283,148
30,889,604,890
IssuesEvent
2023-08-04 02:58:43
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2020-8649 (Medium) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2020-8649 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> There is a use-after-free vulnerability in the Linux kernel through 5.5.2 in the vgacon_invert_region function in drivers/video/console/vgacon.c. <p>Publish Date: 2020-02-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8649>CVE-2020-8649</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-8649">https://www.linuxkernelcves.com/cves/CVE-2020-8649</a></p> <p>Release Date: 2020-02-06</p> <p>Fix Resolution: v5.6-rc5,v3.16.83,v4.14.173,v4.19.109,v4.4.216,v4.9.216,v5.4.25,v5.5.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8649 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2020-8649 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/console/vgacon.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> There is a use-after-free vulnerability in the Linux kernel through 5.5.2 in the vgacon_invert_region function in drivers/video/console/vgacon.c. <p>Publish Date: 2020-02-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8649>CVE-2020-8649</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-8649">https://www.linuxkernelcves.com/cves/CVE-2020-8649</a></p> <p>Release Date: 2020-02-06</p> <p>Fix Resolution: v5.6-rc5,v3.16.83,v4.14.173,v4.19.109,v4.4.216,v4.9.216,v5.4.25,v5.5.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers video console vgacon c drivers video console vgacon c vulnerability details there is a use after free vulnerability in the linux kernel through in the vgacon invert region function in drivers video console vgacon c publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
280,432
21,260,072,498
IssuesEvent
2022-04-13 02:31:42
mikezimm/SecureScript7
https://api.github.com/repos/mikezimm/SecureScript7
closed
Add Debugging code tips in prop pane setup
documentation enhancement Complete
It's hard to debug. Add a tips with lessons learned in the web part.
1.0
Add Debugging code tips in prop pane setup - It's hard to debug. Add a tips with lessons learned in the web part.
non_process
add debugging code tips in prop pane setup it s hard to debug add a tips with lessons learned in the web part
0
87,780
25,209,199,648
IssuesEvent
2022-11-14 01:02:52
OpenBB-finance/OpenBBTerminal
https://api.github.com/repos/OpenBB-finance/OpenBBTerminal
closed
[Bug] Windows Installer Broken with Poetry Install
bug build installer
**Describe the bug** Windows installer is broken. [Link](https://github.com/OpenBB-finance/OpenBBTerminal/actions/runs/3286347083/jobs/5414352853) <img width="1340" alt="Screen Shot 2022-10-20 at 2 10 09 PM" src="https://user-images.githubusercontent.com/53658028/197025636-54437cb9-7ae1-4934-a4e2-96ae52db3a49.png">
1.0
[Bug] Windows Installer Broken with Poetry Install - **Describe the bug** Windows installer is broken. [Link](https://github.com/OpenBB-finance/OpenBBTerminal/actions/runs/3286347083/jobs/5414352853) <img width="1340" alt="Screen Shot 2022-10-20 at 2 10 09 PM" src="https://user-images.githubusercontent.com/53658028/197025636-54437cb9-7ae1-4934-a4e2-96ae52db3a49.png">
non_process
windows installer broken with poetry install describe the bug windows installer is broken img width alt screen shot at pm src
0
10,640
13,446,159,186
IssuesEvent
2020-09-08 12:33:48
MHRA/products
https://api.github.com/repos/MHRA/products
closed
PARS - Design review for search results display
EPIC - PARs process NEW :new:
TASK Design tweaks needed to the search results display, so that they are in line with SPC / PILs ### Acceptance Criteria - [ ] The commas are removed from the between product name, dose, strength etc - [ ] There is a dash between the name and the PL number (example...... Ibuprofen 200mg Tablets (Ibuprofen) - PL 17907/0160) **Customer acceptance criteria** **Technical acceptance criteria** **Data acceptance criteria** **Testing acceptance criteria** **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
PARS - Design review for search results display - TASK Design tweaks needed to the search results display, so that they are in line with SPC / PILs ### Acceptance Criteria - [ ] The commas are removed from the between product name, dose, strength etc - [ ] There is a dash between the name and the PL number (example...... Ibuprofen 200mg Tablets (Ibuprofen) - PL 17907/0160) **Customer acceptance criteria** **Technical acceptance criteria** **Data acceptance criteria** **Testing acceptance criteria** **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
pars design review for search results display task design tweaks needed to the search results display so that they are in line with spc pils acceptance criteria the commas are removed from the between product name dose strength etc there is a dash between the name and the pl number example ibuprofen tablets ibuprofen pl customer acceptance criteria technical acceptance criteria data acceptance criteria testing acceptance criteria size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
1
2,635
5,412,522,136
IssuesEvent
2017-03-01 14:45:38
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
create fortran preprocessor output
enhancement preprocessor
dynamic and static files each as independent functions model as a fortran module
1.0
create fortran preprocessor output - dynamic and static files each as independent functions model as a fortran module
process
create fortran preprocessor output dynamic and static files each as independent functions model as a fortran module
1
93,849
19,346,556,203
IssuesEvent
2021-12-15 11:25:55
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
closed
Short solution needed: "How to list docker tags" (docker)
help wanted good first issue code docker
Please help us write most modern and shortest code solution for this issue: **How to list docker tags** (technology: [docker](https://onelinerhub.com/docker)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to use comments to make solution explained. 3. Link to this issue in comments of pull request.
1.0
Short solution needed: "How to list docker tags" (docker) - Please help us write most modern and shortest code solution for this issue: **How to list docker tags** (technology: [docker](https://onelinerhub.com/docker)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to use comments to make solution explained. 3. Link to this issue in comments of pull request.
non_process
short solution needed how to list docker tags docker please help us write most modern and shortest code solution for this issue how to list docker tags technology fast way just write the code solution in the comments prefered way create pull request with a new code file inside don t forget to use comments to make solution explained link to this issue in comments of pull request
0
249,185
7,953,975,332
IssuesEvent
2018-07-12 05:11:51
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Placed items disappear but block placement
Medium Priority
**Version:** 0.7.4.5 beta **Expected behavior:** When placing an item (so far tried ramp, chair), it should appear. **Actual behavior:** The item disappears immediately. I can walk through it, but I cant place anything else there (blocks or items). The item can be deleted with the Dev Tool (ramps need to be deleted block by block) **Steps to reproduce:** In the savegame, in the area around x=373 y=210 (bothin a tunnel at z=47 and on a mountain at z=56) try placing an item (tested with ramps, chairs), item disappears, placement of items or blocks not possible Invisible item can be deleted with the DevTool, but the same problem occurs again when placing a new item ![20180523192717_1](https://user-images.githubusercontent.com/39566560/40443397-6f311ea0-5ec6-11e8-9669-84f945e5c053.jpg) ![20180523192726_1](https://user-images.githubusercontent.com/39566560/40443401-71678fce-5ec6-11e8-901b-3bdc5d99987b.jpg) ![20180523192745_1](https://user-images.githubusercontent.com/39566560/40443402-732647c4-5ec6-11e8-865e-e714fcb5106d.jpg) ![20180523192750_1](https://user-images.githubusercontent.com/39566560/40443404-749d3cfc-5ec6-11e8-833a-2def2700be9d.jpg)
1.0
Placed items disappear but block placement - **Version:** 0.7.4.5 beta **Expected behavior:** When placing an item (so far tried ramp, chair), it should appear. **Actual behavior:** The item disappears immediately. I can walk through it, but I cant place anything else there (blocks or items). The item can be deleted with the Dev Tool (ramps need to be deleted block by block) **Steps to reproduce:** In the savegame, in the area around x=373 y=210 (bothin a tunnel at z=47 and on a mountain at z=56) try placing an item (tested with ramps, chairs), item disappears, placement of items or blocks not possible Invisible item can be deleted with the DevTool, but the same problem occurs again when placing a new item ![20180523192717_1](https://user-images.githubusercontent.com/39566560/40443397-6f311ea0-5ec6-11e8-9669-84f945e5c053.jpg) ![20180523192726_1](https://user-images.githubusercontent.com/39566560/40443401-71678fce-5ec6-11e8-901b-3bdc5d99987b.jpg) ![20180523192745_1](https://user-images.githubusercontent.com/39566560/40443402-732647c4-5ec6-11e8-865e-e714fcb5106d.jpg) ![20180523192750_1](https://user-images.githubusercontent.com/39566560/40443404-749d3cfc-5ec6-11e8-833a-2def2700be9d.jpg)
non_process
placed items disappear but block placement version beta expected behavior when placing an item so far tried ramp chair it should appear actual behavior the item disappears immediately i can walk through it but i cant place anything else there blocks or items the item can be deleted with the dev tool ramps need to be deleted block by block steps to reproduce in the savegame in the area around x y bothin a tunnel at z and on a mountain at z try placing an item tested with ramps chairs item disappears placement of items or blocks not possible invisible item can be deleted with the devtool but the same problem occurs again when placing a new item
0
20,950
27,809,757,770
IssuesEvent
2023-03-18 01:38:59
serai-dex/serai
https://api.github.com/repos/serai-dex/serai
closed
develop CI is failing
bug CI/CD processor bitcoin
It looks like the Bitcoin node the processor expects isn't being reached/isn't configured as expected?
1.0
develop CI is failing - It looks like the Bitcoin node the processor expects isn't being reached/isn't configured as expected?
process
develop ci is failing it looks like the bitcoin node the processor expects isn t being reached isn t configured as expected
1
9,184
12,228,607,382
IssuesEvent
2020-05-03 20:09:42
chfor183/data_science_articles
https://api.github.com/repos/chfor183/data_science_articles
opened
Imbalanced Datasets
Data Preprocessing
## TL;DR Yes ! ## Key Takeaways - 1 - 2 ## Useful Code Snippets ``` function test() { console.log("notice the blank line before this function?"); } ``` ## Articles/Ressources https://towardsdatascience.com/handling-imbalanced-datasets-in-machine-learning-7a0e84220f28 https://towardsdatascience.com/imbalanced-class-sizes-and-classification-models-a-cautionary-tale-3648b8586e03
1.0
Imbalanced Datasets - ## TL;DR Yes ! ## Key Takeaways - 1 - 2 ## Useful Code Snippets ``` function test() { console.log("notice the blank line before this function?"); } ``` ## Articles/Ressources https://towardsdatascience.com/handling-imbalanced-datasets-in-machine-learning-7a0e84220f28 https://towardsdatascience.com/imbalanced-class-sizes-and-classification-models-a-cautionary-tale-3648b8586e03
process
imbalanced datasets tl dr yes key takeaways useful code snippets function test console log notice the blank line before this function articles ressources
1
35,818
6,499,122,969
IssuesEvent
2017-08-22 20:12:15
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Please note versions in Bazel docs
P2 type: documentation
On TensorFlow and Closure Rules, I want to be able to provide at least some level of support for old versions of Bazel. This is currently not easy to do, because the documentation doesn't tell me at which version a feature was introduced, or became deprecated. For example, [ConfigurationTransition](https://www.bazel.io/versions/master/docs/skylark/lib/ConfigurationTransition.html) doesn't say at which version `"host"` instead of `HOST_CFG` started working. Or the documentation for [alias()](https://www.bazel.io/versions/master/docs/be/general.html#alias) doesn't tell me in which version it was introduced. I understand Bazel isn't quite at 1.0 yet. But I still think taking the extra time to do this would be worthwhile.
1.0
Please note versions in Bazel docs - On TensorFlow and Closure Rules, I want to be able to provide at least some level of support for old versions of Bazel. This is currently not easy to do, because the documentation doesn't tell me at which version a feature was introduced, or became deprecated. For example, [ConfigurationTransition](https://www.bazel.io/versions/master/docs/skylark/lib/ConfigurationTransition.html) doesn't say at which version `"host"` instead of `HOST_CFG` started working. Or the documentation for [alias()](https://www.bazel.io/versions/master/docs/be/general.html#alias) doesn't tell me in which version it was introduced. I understand Bazel isn't quite at 1.0 yet. But I still think taking the extra time to do this would be worthwhile.
non_process
please note versions in bazel docs on tensorflow and closure rules i want to be able to provide at least some level of support for old versions of bazel this is currently not easy to do because the documentation doesn t tell me at which version a feature was introduced or became deprecated for example doesn t say at which version host instead of host cfg started working or the documentation for doesn t tell me in which version it was introduced i understand bazel isn t quite at yet but i still think taking the extra time to do this would be worthwhile
0
205,977
15,702,893,815
IssuesEvent
2021-03-26 13:13:24
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
opened
Feature Test Summary
Feature Test Summary team:Zombie Apocalypse
## Test Strategy Test strategy followed what was done for the addition of a very similar property that allowed the sharing of connections on direct lookups. ### Test strategy * What functionality is new or modified by this feature? Feature adds a boolean attribute `enableContainerAuthForDirectLookups` to existing connectionManager config element. If the value of the new attrib is true, when a datasource is looked up w/o injection or resource-ref, the default authentication mechanism will be CONTAINER instead of APPLICATION. Prior to this feature, the only way to change the autheniticaton was via a resource -ref. * What are the positive and negative tests for that functionality? Since the default behavior is unchanged, ie, APPLICATION auth, and we have tests that perform direct lookups as part of other tests, the negative test path is already covered. For the positive, a test was added to the com.ibm.ws.jdbc_fat test suite that sets the new attrib to true, performs a lookup, obtains a connection and examines the credentials used by the database to make certain the specified container auth user was used to login and not the default. * What manual tests are there (if any)? None. ## Confidence Level **Collectively as a team you need to assess your confidence in the testing delivered based on the values below. This should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole.** Please indicate your confidence in the testing (up to and including FAT) delivered with this feature by selecting one of these values: 0 - No automated testing delivered 1 - We have minimal automated coverage of the feature including golden paths. There is a relatively high risk that defects or issues could be found in this feature. 2 - We have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here. Error/outlying scenarios are not really covered. There are likely risks that issues may exist in the golden paths 3 - We have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error/outlying scenarios. There is a risk when the feature is used outside the golden paths however we are confident on the golden path. Note: This may still be a valid end state for a feature... things like Beta features may well suffice at this level. 4 - We have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error/outlying scenarios. While more testing of the error/outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide. 5 - We have delivered all automated testing we believe is needed for this feature. The testing covers all golden path cases as well as all the error/outlying scenarios that make sense. We are not aware of any gaps in the testing at this time. No manual testing is required to verify this feature. Based on your answer above, for any answer other than a 4 or 5 please provide details of what drove your answer. Please be aware, it may be perfectly reasonable in some scenarios to deliver with any value above. We may accept no automated testing is needed for some features, we may be happy with low levels of testing on samples for instance so please don't feel the need to drive to a 5. We need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid. What are the gaps, what is the risk etc. Please also provide links to the follow on work that is needed to close the gaps (should you deem it needed)
1.0
Feature Test Summary - ## Test Strategy Test strategy followed what was done for the addition of a very similar property that allowed the sharing of connections on direct lookups. ### Test strategy * What functionality is new or modified by this feature? Feature adds a boolean attribute `enableContainerAuthForDirectLookups` to existing connectionManager config element. If the value of the new attrib is true, when a datasource is looked up w/o injection or resource-ref, the default authentication mechanism will be CONTAINER instead of APPLICATION. Prior to this feature, the only way to change the autheniticaton was via a resource -ref. * What are the positive and negative tests for that functionality? Since the default behavior is unchanged, ie, APPLICATION auth, and we have tests that perform direct lookups as part of other tests, the negative test path is already covered. For the positive, a test was added to the com.ibm.ws.jdbc_fat test suite that sets the new attrib to true, performs a lookup, obtains a connection and examines the credentials used by the database to make certain the specified container auth user was used to login and not the default. * What manual tests are there (if any)? None. ## Confidence Level **Collectively as a team you need to assess your confidence in the testing delivered based on the values below. This should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole.** Please indicate your confidence in the testing (up to and including FAT) delivered with this feature by selecting one of these values: 0 - No automated testing delivered 1 - We have minimal automated coverage of the feature including golden paths. There is a relatively high risk that defects or issues could be found in this feature. 2 - We have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here. Error/outlying scenarios are not really covered. There are likely risks that issues may exist in the golden paths 3 - We have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error/outlying scenarios. There is a risk when the feature is used outside the golden paths however we are confident on the golden path. Note: This may still be a valid end state for a feature... things like Beta features may well suffice at this level. 4 - We have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error/outlying scenarios. While more testing of the error/outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide. 5 - We have delivered all automated testing we believe is needed for this feature. The testing covers all golden path cases as well as all the error/outlying scenarios that make sense. We are not aware of any gaps in the testing at this time. No manual testing is required to verify this feature. Based on your answer above, for any answer other than a 4 or 5 please provide details of what drove your answer. Please be aware, it may be perfectly reasonable in some scenarios to deliver with any value above. We may accept no automated testing is needed for some features, we may be happy with low levels of testing on samples for instance so please don't feel the need to drive to a 5. We need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid. What are the gaps, what is the risk etc. Please also provide links to the follow on work that is needed to close the gaps (should you deem it needed)
non_process
feature test summary test strategy test strategy followed what was done for the addition of a very similar property that allowed the sharing of connections on direct lookups test strategy what functionality is new or modified by this feature feature adds a boolean attribute enablecontainerauthfordirectlookups to existing connectionmanager config element if the value of the new attrib is true when a datasource is looked up w o injection or resource ref the default authentication mechanism will be container instead of application prior to this feature the only way to change the autheniticaton was via a resource ref what are the positive and negative tests for that functionality since the default behavior is unchanged ie application auth and we have tests that perform direct lookups as part of other tests the negative test path is already covered for the positive a test was added to the com ibm ws jdbc fat test suite that sets the new attrib to true performs a lookup obtains a connection and examines the credentials used by the database to make certain the specified container auth user was used to login and not the default what manual tests are there if any none confidence level collectively as a team you need to assess your confidence in the testing delivered based on the values below this should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole please indicate your confidence in the testing up to and including fat delivered with this feature by selecting one of these values no automated testing delivered we have minimal automated coverage of the feature including golden paths there is a relatively high risk that defects or issues could be found in this feature we have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here error outlying scenarios are not really covered there are likely risks that issues may exist in the golden paths we have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error outlying scenarios there is a risk when the feature is used outside the golden paths however we are confident on the golden path note this may still be a valid end state for a feature things like beta features may well suffice at this level we have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error outlying scenarios while more testing of the error outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide we have delivered all automated testing we believe is needed for this feature the testing covers all golden path cases as well as all the error outlying scenarios that make sense we are not aware of any gaps in the testing at this time no manual testing is required to verify this feature based on your answer above for any answer other than a or please provide details of what drove your answer please be aware it may be perfectly reasonable in some scenarios to deliver with any value above we may accept no automated testing is needed for some features we may be happy with low levels of testing on samples for instance so please don t feel the need to drive to a we need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid what are the gaps what is the risk etc please also provide links to the follow on work that is needed to close the gaps should you deem it needed
0
15,747
19,911,552,479
IssuesEvent
2022-01-25 17:39:54
input-output-hk/high-assurance-legacy
https://api.github.com/repos/input-output-hk/high-assurance-legacy
closed
Add Isabelle export of Haskell-embedded ♮-calculus code
type: enhancement language: isabelle language: haskell topic: process calculus
Our goal is to implement an interpretation that turns a ♮-calculus program into Isabelle source code that represents this program using the Isabelle embedding of the calculus.
1.0
Add Isabelle export of Haskell-embedded ♮-calculus code - Our goal is to implement an interpretation that turns a ♮-calculus program into Isabelle source code that represents this program using the Isabelle embedding of the calculus.
process
add isabelle export of haskell embedded ♮ calculus code our goal is to implement an interpretation that turns a ♮ calculus program into isabelle source code that represents this program using the isabelle embedding of the calculus
1
18,801
24,702,423,445
IssuesEvent
2022-10-19 16:13:41
NREL/buildstockbatch
https://api.github.com/repos/NREL/buildstockbatch
closed
Timeseries processing doesn't get triggered for ComStock projects
bug postprocessing eagle
**Describe the bug** The timeseries postprocessing isn't triggered in ComStock due to misalignment of the workflow generators for commercial and residential (see #222). Basically https://github.com/NREL/buildstockbatch/blob/1cca3c82fba1d013f7f81bbad4ba05a4f6521d11/buildstockbatch/base.py#L545 is not triggered by the commercial workflow yaml which looks like this: ```yml workflow_generator: type: commercial_default args: reporting_measures: - measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results arguments: building_summary_section: true annual_overview_section: true monthly_overview_section: true utility_bills_rates_section: true envelope_section_section: true space_type_breakdown_section: true space_type_details_section: true interior_lighting_section: true plug_loads_section: true exterior_light_section: true water_use_section: true hvac_load_profile: true zone_condition_section: true zone_summary_section: true zone_equipment_detail_section: true air_loops_detail_section: true plant_loops_detail_section: true outdoor_air_section: true cost_summary_section: true source_energy_section: true schedules_overview_section: true - measure_dir_name: SimulationOutputReport - measure_dir_name: TimeseriesCSVExport arguments: reporting_frequency: "Timestep" inc_output_variables: false - measure_dir_name: comstock_sensitivity_reports - measure_dir_name: qoi_report - measure_dir_name: la_100_qaqc arguments: run_qaqc: false - measure_dir_name: simulation_settings_check arguments: run_sim_settings_checks: true ``` **To Reproduce** Steps to reproduce the behavior: 1. Example yaml from @mdahlhausen: ```yml schema_version: '0.3' buildstock_directory: /lustre/eaglefs/projects/enduse/comstock/repos/comstock project_directory: national output_directory: /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/ weather_files_path: /lustre/eaglefs/projects/enduse/comstock/weather/Fort_Collins_County_IDs_2016.zip sampler: type: precomputed args: sample_file: /lustre/eaglefs/projects/enduse/comstock/samples/buildstock_wwr_2016_test.csv eagle: account: enduse n_jobs: 1 minutes_per_sim: 120 postprocessing: time: 45 postprocessing: keep_individual_timeseries: true aws: region_name: 'us-west-2' s3: bucket: eulp prefix: simulation_output/regional_runs/comstock/com_test_17_2016/ athena: glue_service_role: service-role/AWSGlueServiceRole-default database_name: enduse max_crawling_time: 600 # Time to wait for the crawler to complete before aborting it baseline: n_buildings_represented: 30 custom_gems: True os_version: eulp_com_v12 os_sha: 3472e8b799 workflow_generator: type: commercial_default args: reporting_measures: - measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results arguments: building_summary_section: true annual_overview_section: true monthly_overview_section: true utility_bills_rates_section: true envelope_section_section: true space_type_breakdown_section: true space_type_details_section: true interior_lighting_section: true plug_loads_section: true exterior_light_section: true water_use_section: true hvac_load_profile: true zone_condition_section: true zone_summary_section: true zone_equipment_detail_section: true air_loops_detail_section: true plant_loops_detail_section: true outdoor_air_section: true cost_summary_section: true source_energy_section: true schedules_overview_section: true - measure_dir_name: SimulationOutputReport - measure_dir_name: TimeseriesCSVExport arguments: reporting_frequency: "Timestep" inc_output_variables: false - measure_dir_name: comstock_sensitivity_reports - measure_dir_name: qoi_report - measure_dir_name: la_100_qaqc arguments: run_qaqc: false - measure_dir_name: simulation_settings_check arguments: run_sim_settings_checks: true ``` 2. That's about it. **Expected behavior** I want (but don't expect given the code) for the timeseries postprocessing to run. But it doesn't. **Logs** postprocessing.out ``` begin eagle_postprocessing.sh Job ID: 6997691 Hostname: r2i7n29 UPLOADONLY: False MEMORY: 85248 head node r2i7n29 workers r8i0n[11,15] r8i0n11: Warning: Permanently added 'r8i0n11,10.148.8.126' (ECDSA) to the list of known hosts. r8i0n15: Warning: Permanently added 'r8i0n15,10.148.8.134' (ECDSA) to the list of known hosts. r8i0n15: total used free shared buff/cache available r8i0n15: Mem: 92G 2.9G 84G 5.1G 5.2G 84G r8i0n15: Swap: 0B 0B 0B r8i0n11: total used free shared buff/cache available r8i0n11: Mem: 92G 2.9G 85G 3.9G 4.1G 85G r8i0n11: Swap: 0B 0B 0B DEBUG:2021-06-15 13:01:24:buildstockbatch.base:Using OpenStudio version: eulp_com_v12 with SHA: 3472e8b799 DEBUG:2021-06-15 13:01:24:__main__:Output directory = /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016 DEBUG:2021-06-15 13:01:24:__main__:Starting postprocessing INFO:2021-06-15 13:01:28:buildstockbatch.postprocessing:Creating results_df. INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Writing /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/results/results_csvs/results_up00.csv.gz INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Uploading the parquet files to s3 INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Upload to S3 completed. The files are uploaded to: eulp/simulation_output/regional_runs/comstock/com_test_17_2016/com_test_17_2016/ INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Creating Athena tables using glue crawler INFO:2021-06-15 13:01:39:buildstockbatch.postprocessing:Deleting existing tables in db enduse: ['_com_test_17_2016']. And creating new ones. INFO:2021-06-15 13:01:51:buildstockbatch.postprocessing:Crawler started INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:Crawler has completed running. It is STOPPING. INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:TablesCreated: 1 TablesUpdated: 0 TablesDeleted: 0 INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Crawler stopped. Deleting it now. INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Removing results_job*.json.gz real 5m45.335s user 0m1.451s sys 0m2.137s ``` **Platform (please complete the following information):** - Simulation platform: eagle - BuildStockBatch version, branch, or sha: 1cca3c82fba1d013f7f81bbad4ba05a4f6521d11 - resstock or comstock repo version, branch, or sha: no clue - Local Desktop OS: n/a **Additional context** This is really a issue with the workflow generator schema / specification but I have no idea what the proper solution is there. Instead i'm going to role a crappy patch for this that will get us through. In general I'm not a huge fan of how timeseries postprocessing is (or really isn't) specified in the yaml but that's also a different issue for a different day.
1.0
Timeseries processing doesn't get triggered for ComStock projects - **Describe the bug** The timeseries postprocessing isn't triggered in ComStock due to misalignment of the workflow generators for commercial and residential (see #222). Basically https://github.com/NREL/buildstockbatch/blob/1cca3c82fba1d013f7f81bbad4ba05a4f6521d11/buildstockbatch/base.py#L545 is not triggered by the commercial workflow yaml which looks like this: ```yml workflow_generator: type: commercial_default args: reporting_measures: - measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results arguments: building_summary_section: true annual_overview_section: true monthly_overview_section: true utility_bills_rates_section: true envelope_section_section: true space_type_breakdown_section: true space_type_details_section: true interior_lighting_section: true plug_loads_section: true exterior_light_section: true water_use_section: true hvac_load_profile: true zone_condition_section: true zone_summary_section: true zone_equipment_detail_section: true air_loops_detail_section: true plant_loops_detail_section: true outdoor_air_section: true cost_summary_section: true source_energy_section: true schedules_overview_section: true - measure_dir_name: SimulationOutputReport - measure_dir_name: TimeseriesCSVExport arguments: reporting_frequency: "Timestep" inc_output_variables: false - measure_dir_name: comstock_sensitivity_reports - measure_dir_name: qoi_report - measure_dir_name: la_100_qaqc arguments: run_qaqc: false - measure_dir_name: simulation_settings_check arguments: run_sim_settings_checks: true ``` **To Reproduce** Steps to reproduce the behavior: 1. Example yaml from @mdahlhausen: ```yml schema_version: '0.3' buildstock_directory: /lustre/eaglefs/projects/enduse/comstock/repos/comstock project_directory: national output_directory: /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/ weather_files_path: /lustre/eaglefs/projects/enduse/comstock/weather/Fort_Collins_County_IDs_2016.zip sampler: type: precomputed args: sample_file: /lustre/eaglefs/projects/enduse/comstock/samples/buildstock_wwr_2016_test.csv eagle: account: enduse n_jobs: 1 minutes_per_sim: 120 postprocessing: time: 45 postprocessing: keep_individual_timeseries: true aws: region_name: 'us-west-2' s3: bucket: eulp prefix: simulation_output/regional_runs/comstock/com_test_17_2016/ athena: glue_service_role: service-role/AWSGlueServiceRole-default database_name: enduse max_crawling_time: 600 # Time to wait for the crawler to complete before aborting it baseline: n_buildings_represented: 30 custom_gems: True os_version: eulp_com_v12 os_sha: 3472e8b799 workflow_generator: type: commercial_default args: reporting_measures: - measure_dir_name: f8e23017-894d-4bdf-977f-37e3961e6f42 # OpenStudio Results arguments: building_summary_section: true annual_overview_section: true monthly_overview_section: true utility_bills_rates_section: true envelope_section_section: true space_type_breakdown_section: true space_type_details_section: true interior_lighting_section: true plug_loads_section: true exterior_light_section: true water_use_section: true hvac_load_profile: true zone_condition_section: true zone_summary_section: true zone_equipment_detail_section: true air_loops_detail_section: true plant_loops_detail_section: true outdoor_air_section: true cost_summary_section: true source_energy_section: true schedules_overview_section: true - measure_dir_name: SimulationOutputReport - measure_dir_name: TimeseriesCSVExport arguments: reporting_frequency: "Timestep" inc_output_variables: false - measure_dir_name: comstock_sensitivity_reports - measure_dir_name: qoi_report - measure_dir_name: la_100_qaqc arguments: run_qaqc: false - measure_dir_name: simulation_settings_check arguments: run_sim_settings_checks: true ``` 2. That's about it. **Expected behavior** I want (but don't expect given the code) for the timeseries postprocessing to run. But it doesn't. **Logs** postprocessing.out ``` begin eagle_postprocessing.sh Job ID: 6997691 Hostname: r2i7n29 UPLOADONLY: False MEMORY: 85248 head node r2i7n29 workers r8i0n[11,15] r8i0n11: Warning: Permanently added 'r8i0n11,10.148.8.126' (ECDSA) to the list of known hosts. r8i0n15: Warning: Permanently added 'r8i0n15,10.148.8.134' (ECDSA) to the list of known hosts. r8i0n15: total used free shared buff/cache available r8i0n15: Mem: 92G 2.9G 84G 5.1G 5.2G 84G r8i0n15: Swap: 0B 0B 0B r8i0n11: total used free shared buff/cache available r8i0n11: Mem: 92G 2.9G 85G 3.9G 4.1G 85G r8i0n11: Swap: 0B 0B 0B DEBUG:2021-06-15 13:01:24:buildstockbatch.base:Using OpenStudio version: eulp_com_v12 with SHA: 3472e8b799 DEBUG:2021-06-15 13:01:24:__main__:Output directory = /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016 DEBUG:2021-06-15 13:01:24:__main__:Starting postprocessing INFO:2021-06-15 13:01:28:buildstockbatch.postprocessing:Creating results_df. INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Writing /lustre/eaglefs/projects/enduse/comstock/com_test_17_2016/results/results_csvs/results_up00.csv.gz INFO:2021-06-15 13:01:30:buildstockbatch.postprocessing:Uploading the parquet files to s3 INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Upload to S3 completed. The files are uploaded to: eulp/simulation_output/regional_runs/comstock/com_test_17_2016/com_test_17_2016/ INFO:2021-06-15 13:01:38:buildstockbatch.postprocessing:Creating Athena tables using glue crawler INFO:2021-06-15 13:01:39:buildstockbatch.postprocessing:Deleting existing tables in db enduse: ['_com_test_17_2016']. And creating new ones. INFO:2021-06-15 13:01:51:buildstockbatch.postprocessing:Crawler started INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:Crawler has completed running. It is STOPPING. INFO:2021-06-15 13:03:22:buildstockbatch.postprocessing:TablesCreated: 1 TablesUpdated: 0 TablesDeleted: 0 INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Crawler stopped. Deleting it now. INFO:2021-06-15 13:06:54:buildstockbatch.postprocessing:Removing results_job*.json.gz real 5m45.335s user 0m1.451s sys 0m2.137s ``` **Platform (please complete the following information):** - Simulation platform: eagle - BuildStockBatch version, branch, or sha: 1cca3c82fba1d013f7f81bbad4ba05a4f6521d11 - resstock or comstock repo version, branch, or sha: no clue - Local Desktop OS: n/a **Additional context** This is really a issue with the workflow generator schema / specification but I have no idea what the proper solution is there. Instead i'm going to role a crappy patch for this that will get us through. In general I'm not a huge fan of how timeseries postprocessing is (or really isn't) specified in the yaml but that's also a different issue for a different day.
process
timeseries processing doesn t get triggered for comstock projects describe the bug the timeseries postprocessing isn t triggered in comstock due to misalignment of the workflow generators for commercial and residential see basically is not triggered by the commercial workflow yaml which looks like this yml workflow generator type commercial default args reporting measures measure dir name openstudio results arguments building summary section true annual overview section true monthly overview section true utility bills rates section true envelope section section true space type breakdown section true space type details section true interior lighting section true plug loads section true exterior light section true water use section true hvac load profile true zone condition section true zone summary section true zone equipment detail section true air loops detail section true plant loops detail section true outdoor air section true cost summary section true source energy section true schedules overview section true measure dir name simulationoutputreport measure dir name timeseriescsvexport arguments reporting frequency timestep inc output variables false measure dir name comstock sensitivity reports measure dir name qoi report measure dir name la qaqc arguments run qaqc false measure dir name simulation settings check arguments run sim settings checks true to reproduce steps to reproduce the behavior example yaml from mdahlhausen yml schema version buildstock directory lustre eaglefs projects enduse comstock repos comstock project directory national output directory lustre eaglefs projects enduse comstock com test weather files path lustre eaglefs projects enduse comstock weather fort collins county ids zip sampler type precomputed args sample file lustre eaglefs projects enduse comstock samples buildstock wwr test csv eagle account enduse n jobs minutes per sim postprocessing time postprocessing keep individual timeseries true aws region name us west bucket eulp prefix simulation output regional runs comstock com test athena glue service role service role awsglueservicerole default database name enduse max crawling time time to wait for the crawler to complete before aborting it baseline n buildings represented custom gems true os version eulp com os sha workflow generator type commercial default args reporting measures measure dir name openstudio results arguments building summary section true annual overview section true monthly overview section true utility bills rates section true envelope section section true space type breakdown section true space type details section true interior lighting section true plug loads section true exterior light section true water use section true hvac load profile true zone condition section true zone summary section true zone equipment detail section true air loops detail section true plant loops detail section true outdoor air section true cost summary section true source energy section true schedules overview section true measure dir name simulationoutputreport measure dir name timeseriescsvexport arguments reporting frequency timestep inc output variables false measure dir name comstock sensitivity reports measure dir name qoi report measure dir name la qaqc arguments run qaqc false measure dir name simulation settings check arguments run sim settings checks true that s about it expected behavior i want but don t expect given the code for the timeseries postprocessing to run but it doesn t logs postprocessing out begin eagle postprocessing sh job id hostname uploadonly false memory head node workers warning permanently added ecdsa to the list of known hosts warning permanently added ecdsa to the list of known hosts total used free shared buff cache available mem swap total used free shared buff cache available mem swap debug buildstockbatch base using openstudio version eulp com with sha debug main output directory lustre eaglefs projects enduse comstock com test debug main starting postprocessing info buildstockbatch postprocessing creating results df info buildstockbatch postprocessing writing lustre eaglefs projects enduse comstock com test results results csvs results csv gz info buildstockbatch postprocessing uploading the parquet files to info buildstockbatch postprocessing upload to completed the files are uploaded to eulp simulation output regional runs comstock com test com test info buildstockbatch postprocessing creating athena tables using glue crawler info buildstockbatch postprocessing deleting existing tables in db enduse and creating new ones info buildstockbatch postprocessing crawler started info buildstockbatch postprocessing crawler has completed running it is stopping info buildstockbatch postprocessing tablescreated tablesupdated tablesdeleted info buildstockbatch postprocessing crawler stopped deleting it now info buildstockbatch postprocessing removing results job json gz real user sys platform please complete the following information simulation platform eagle buildstockbatch version branch or sha resstock or comstock repo version branch or sha no clue local desktop os n a additional context this is really a issue with the workflow generator schema specification but i have no idea what the proper solution is there instead i m going to role a crappy patch for this that will get us through in general i m not a huge fan of how timeseries postprocessing is or really isn t specified in the yaml but that s also a different issue for a different day
1
92,717
8,376,569,481
IssuesEvent
2018-10-05 20:18:27
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
[GUI] Map view - centralization
Priority:High Status:ToTest Team:Frontend Type:Bug
Map centralized in the region of Pradópolis but devices are in Campinas. ![screenshot_31](https://user-images.githubusercontent.com/37310063/45839994-37d7e680-bcec-11e8-8533-5b44db642a37.png) ![screenshot_30](https://user-images.githubusercontent.com/37310063/45839980-2db5e800-bcec-11e8-90e9-5e59b74e7f20.png) **Affected Version:** 0.3.0-nightly_20180912
1.0
[GUI] Map view - centralization - Map centralized in the region of Pradópolis but devices are in Campinas. ![screenshot_31](https://user-images.githubusercontent.com/37310063/45839994-37d7e680-bcec-11e8-8533-5b44db642a37.png) ![screenshot_30](https://user-images.githubusercontent.com/37310063/45839980-2db5e800-bcec-11e8-90e9-5e59b74e7f20.png) **Affected Version:** 0.3.0-nightly_20180912
non_process
map view centralization map centralized in the region of pradópolis but devices are in campinas affected version nightly
0
422,205
28,377,116,692
IssuesEvent
2023-04-12 21:55:22
EECE3093C/team-project-weather
https://api.github.com/repos/EECE3093C/team-project-weather
opened
Label all iteration 4 issues as functional/non-functional
documentation
Ensure all issues are labeled as such. This is a functional requirement.
1.0
Label all iteration 4 issues as functional/non-functional - Ensure all issues are labeled as such. This is a functional requirement.
non_process
label all iteration issues as functional non functional ensure all issues are labeled as such this is a functional requirement
0
136,978
11,094,017,108
IssuesEvent
2019-12-16 04:30:53
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
CIS Scan - Failed nodes not reported in the API when the status of test is "mixed"
[zube]: To Test area/scan-tool kind/bug-qa team/ca
**What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - On a custom cluster - all roles - 1 node, run CIS Scans. - For a test which has the state as "mixed", API report does not list the nodes it has failed on **Expected Result:** The report in API should report the nodes which have failed when the state of the test is "mixed" **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master - head - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): custom - Kubernetes version (use `kubectl version`): ``` 1.16.3 ```
1.0
CIS Scan - Failed nodes not reported in the API when the status of test is "mixed" - **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - On a custom cluster - all roles - 1 node, run CIS Scans. - For a test which has the state as "mixed", API report does not list the nodes it has failed on **Expected Result:** The report in API should report the nodes which have failed when the state of the test is "mixed" **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master - head - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): custom - Kubernetes version (use `kubectl version`): ``` 1.16.3 ```
non_process
cis scan failed nodes not reported in the api when the status of test is mixed what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible on a custom cluster all roles node run cis scans for a test which has the state as mixed api report does not list the nodes it has failed on expected result the report in api should report the nodes which have failed when the state of the test is mixed other details that may be helpful environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui master head installation option single install ha single if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported custom kubernetes version use kubectl version
0
17,207
22,792,178,938
IssuesEvent
2022-07-10 06:52:42
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
opened
[processor/metricstransform] Remove empty metrics
priority:p3 processor/metricstransform
Some operations, e.g. `remove_label_values`, can leave metric without data points. That kind of metrics can be considered invalid and should be removed
1.0
[processor/metricstransform] Remove empty metrics - Some operations, e.g. `remove_label_values`, can leave metric without data points. That kind of metrics can be considered invalid and should be removed
process
remove empty metrics some operations e g remove label values can leave metric without data points that kind of metrics can be considered invalid and should be removed
1
47,018
2,971,267,420
IssuesEvent
2015-07-14 04:57:56
nabu-catalog/nabu
https://api.github.com/repos/nabu-catalog/nabu
closed
Delete items and collections
2015 Priorities Priority
We need the ability to delete collections even if they have content, and also to delete items if they have content. At the moment we need to delete each essence file before we can delete an item and each item before we can delete a collection.
1.0
Delete items and collections - We need the ability to delete collections even if they have content, and also to delete items if they have content. At the moment we need to delete each essence file before we can delete an item and each item before we can delete a collection.
non_process
delete items and collections we need the ability to delete collections even if they have content and also to delete items if they have content at the moment we need to delete each essence file before we can delete an item and each item before we can delete a collection
0
1,430
3,995,445,257
IssuesEvent
2016-05-10 15:32:35
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
opened
Wrong pass `this` for function constructor
AREA: client SYSTEM: resource processing TYPE: bug
Found in https://testcafe-hhhm.devexpress.com/history?timestamp=1462872434105&siteUrl=http://shareasale.com/ Script for reproduce: ```javascript var Ext = { toString: function () { return 'Ext'; } }; Ext.data = { toString: function () { return 'Ext.data'; } }; Ext.data.Connection = function () { console.log(this); //Without proxing `this` equal Ext.data.Connection //With proxing `this` equal the parent scope - Ext.data }; Ext.data.Connection.toString = function () { return 'Ext.data.Connection'; } var obj = new Ext.data.Connection(); ``` Possible cause - https://github.com/DevExpress/testcafe-hammerhead/pull/562
1.0
Wrong pass `this` for function constructor - Found in https://testcafe-hhhm.devexpress.com/history?timestamp=1462872434105&siteUrl=http://shareasale.com/ Script for reproduce: ```javascript var Ext = { toString: function () { return 'Ext'; } }; Ext.data = { toString: function () { return 'Ext.data'; } }; Ext.data.Connection = function () { console.log(this); //Without proxing `this` equal Ext.data.Connection //With proxing `this` equal the parent scope - Ext.data }; Ext.data.Connection.toString = function () { return 'Ext.data.Connection'; } var obj = new Ext.data.Connection(); ``` Possible cause - https://github.com/DevExpress/testcafe-hammerhead/pull/562
process
wrong pass this for function constructor found in script for reproduce javascript var ext tostring function return ext ext data tostring function return ext data ext data connection function console log this without proxing this equal ext data connection with proxing this equal the parent scope ext data ext data connection tostring function return ext data connection var obj new ext data connection possible cause
1
22,586
31,811,331,265
IssuesEvent
2023-09-13 17:03:10
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
reopened
Obsoletion request: RNA polyadenylation
RNA processes obsoletion ready MF_in_BP
Please provide as much information as you can: Once https://github.com/geneontology/go-ontology/issues/24921 is done we can begin to obsolete - [x] GO:0043631   RNA polyadenylation - [x] GO:0006378 mRNA polyadenylation - [x] GO:1900364 negative regulation of mRNA polyadenylation - [x] GO:1900365 positive regulation of mRNA polyadenylation - [x] GO:1900363 regulation of mRNA polyadenylation - [x] GO:0043629 ncRNA polyadenylation (I will make sure there are direct replacements for all of these) - [x] GO:0043630 ncRNA polyadenylation involved in polyadenylation-dependent ncRNA catabolic process replace by -> GO:0043634 polyadenylation-dependent ncRNA catabolic process - [x] GO:0071050 sno(s)RNA polyadenylation -> replace by GO:0071051 polyadenylation-dependent snoRNA 3'-end processing Rehouse: - [ ] GO:0097222 mitochondrial mRNA polyadenylation - [ ] GO:0140235 RNA polyadenylation at postsynapse (this could probably move under cytoplasmic polyadenylation, need to check) - [ ] GO:0039698 polyadenylation of viral mRNA by polymerase stuttering (this is something completely different nd ned rehoming) * **GO term ID and Label** * **Reason for deprecation** Put an x in the appropriate box: - [x] The reason for obsoletion is that this term represents a molecular function. * **"Replace by" term (ID and label)** If all annotations can safely be moved to that term * **"Consider" term(s) (ID and label)** Suggestions for reannotation * **Are there annotations to this term?** - How many EXP: * **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)** * **Is this term in a subset? (check the AmiGO page for that term)** * **Any other information** ---- Checklist for ontology editor ***Check term usage and metadata in Protégé*** - [ ] check term usage in the ontology - [ ] check internal mappings: RHEA, EC, MetaCyc - [ ] check subset usage - [ ] check taxon constraints ***Check annotations*** - [x] create Annotation Review spreadsheet https://docs.google.com/spreadsheets/d/1Dyb4jkJAqnHonfWw2-9u3zQcpAcVd5cd9p_I7ta2Muc/edit#gid=0 - [x] Annotation review ticket https://github.com/geneontology/go-annotation/issues/4553 - [x] find external mappings (via IEAS), include in obsoletion notice - [x] tag curators in Annotation Review ticket. Responsible annotator for each group: https://github.com/geneontology/go-site/blob/master/metadata/group-contacts.csv ***Notification*** - [ ] create [obsoletion announcement](https://github.com/geneontology/go-announcements/issues/new?assignees=&labels=obsoletion&template=obsoletion-notice.md&title=Obsoletion+notice%3A+%5BGO+ID%3A+term+label%5D) - [ ] announce to GO friends (go-friends@mailman.stanford.edu) - [ ] paste the text in the ontology ticket
1.0
Obsoletion request: RNA polyadenylation - Please provide as much information as you can: Once https://github.com/geneontology/go-ontology/issues/24921 is done we can begin to obsolete - [x] GO:0043631   RNA polyadenylation - [x] GO:0006378 mRNA polyadenylation - [x] GO:1900364 negative regulation of mRNA polyadenylation - [x] GO:1900365 positive regulation of mRNA polyadenylation - [x] GO:1900363 regulation of mRNA polyadenylation - [x] GO:0043629 ncRNA polyadenylation (I will make sure there are direct replacements for all of these) - [x] GO:0043630 ncRNA polyadenylation involved in polyadenylation-dependent ncRNA catabolic process replace by -> GO:0043634 polyadenylation-dependent ncRNA catabolic process - [x] GO:0071050 sno(s)RNA polyadenylation -> replace by GO:0071051 polyadenylation-dependent snoRNA 3'-end processing Rehouse: - [ ] GO:0097222 mitochondrial mRNA polyadenylation - [ ] GO:0140235 RNA polyadenylation at postsynapse (this could probably move under cytoplasmic polyadenylation, need to check) - [ ] GO:0039698 polyadenylation of viral mRNA by polymerase stuttering (this is something completely different nd ned rehoming) * **GO term ID and Label** * **Reason for deprecation** Put an x in the appropriate box: - [x] The reason for obsoletion is that this term represents a molecular function. * **"Replace by" term (ID and label)** If all annotations can safely be moved to that term * **"Consider" term(s) (ID and label)** Suggestions for reannotation * **Are there annotations to this term?** - How many EXP: * **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)** * **Is this term in a subset? (check the AmiGO page for that term)** * **Any other information** ---- Checklist for ontology editor ***Check term usage and metadata in Protégé*** - [ ] check term usage in the ontology - [ ] check internal mappings: RHEA, EC, MetaCyc - [ ] check subset usage - [ ] check taxon constraints ***Check annotations*** - [x] create Annotation Review spreadsheet https://docs.google.com/spreadsheets/d/1Dyb4jkJAqnHonfWw2-9u3zQcpAcVd5cd9p_I7ta2Muc/edit#gid=0 - [x] Annotation review ticket https://github.com/geneontology/go-annotation/issues/4553 - [x] find external mappings (via IEAS), include in obsoletion notice - [x] tag curators in Annotation Review ticket. Responsible annotator for each group: https://github.com/geneontology/go-site/blob/master/metadata/group-contacts.csv ***Notification*** - [ ] create [obsoletion announcement](https://github.com/geneontology/go-announcements/issues/new?assignees=&labels=obsoletion&template=obsoletion-notice.md&title=Obsoletion+notice%3A+%5BGO+ID%3A+term+label%5D) - [ ] announce to GO friends (go-friends@mailman.stanford.edu) - [ ] paste the text in the ontology ticket
process
obsoletion request rna polyadenylation please provide as much information as you can once is done we can begin to obsolete go    rna polyadenylation go mrna polyadenylation go negative regulation of mrna polyadenylation go positive regulation of mrna polyadenylation go regulation of mrna polyadenylation go ncrna polyadenylation i will make sure there are direct replacements for all of these go ncrna polyadenylation involved in polyadenylation dependent ncrna catabolic process replace by go polyadenylation dependent ncrna catabolic process go sno s rna polyadenylation replace by go polyadenylation dependent snorna end processing rehouse go mitochondrial mrna polyadenylation go rna polyadenylation at postsynapse this could probably move under cytoplasmic polyadenylation need to check go polyadenylation of viral mrna by polymerase stuttering this is something completely different nd ned rehoming go term id and label reason for deprecation put an x in the appropriate box the reason for obsoletion is that this term represents a molecular function replace by term id and label if all annotations can safely be moved to that term consider term s id and label suggestions for reannotation are there annotations to this term how many exp are there mappings and cross references to this term interpro keywords check quickgo cross references section is this term in a subset check the amigo page for that term any other information checklist for ontology editor check term usage and metadata in protégé check term usage in the ontology check internal mappings rhea ec metacyc check subset usage check taxon constraints check annotations create annotation review spreadsheet annotation review ticket find external mappings via ieas include in obsoletion notice tag curators in annotation review ticket responsible annotator for each group notification create announce to go friends go friends mailman stanford edu paste the text in the ontology ticket
1
18,943
24,904,581,998
IssuesEvent
2022-10-29 04:28:53
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
Windows Process.Name() getFromSnapProcess() performance
os:windows package:process performance
Currently `Process.Name()` on Windows calls the `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` syscall to get it's ppid and name: https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L321-L322 https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L830-L831 It seems `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` will return all processes in the system despite passing the ProcessID parameter. Although returning a snapshot of all processes is quite fast most of the time (<10ms), it can sometimes be slow (>100ms, sometimes even >1000ms). Can we change the flags like below according to the documentation: (not so sure about the exact flags, especially `TH32CS_SNAPMODULE32`) ``` func getFromSnapProcess(pid int32) (int32, int32, string, error) { // https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot // To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero. // Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again, // specifying its process identifier and the TH32CS_SNAPHEAPLIST or TH32_SNAPMODULE value. snap, err := windows.CreateToolhelp32Snapshot(windows.TH32CS_SNAPHEAPLIST|windows.TH32CS_SNAPMODULE|windows.TH32CS_SNAPMODULE32, uint32(pid)) if err != nil { return 0, 0, "", err } defer windows.CloseHandle(snap) var pe32 windows.ProcessEntry32 pe32.Size = uint32(unsafe.Sizeof(pe32)) if err = windows.Process32First(snap, &pe32); err != nil && err != syscall.ERROR_NO_MORE_FILES { return 0, 0, "", err } for { if pe32.ProcessID == uint32(pid) { szexe := windows.UTF16ToString(pe32.ExeFile[:]) return int32(pe32.ParentProcessID), int32(pe32.Threads), szexe, nil } if err = windows.Process32Next(snap, &pe32); err != nil { break } } return 0, 0, "", fmt.Errorf("couldn't find pid: %d", pid) } ``` or switch to `NtQuerySystemInformation`? ## refs https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot: - TH32CS_SNAPPROCESS 0x00000002 | Includes all processes in the system in the snapshot. > To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero. > Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again, specifying its process identifier and the **TH32CS_SNAPHEAPLIST** or **TH32_SNAPMODULE** value. ## see also - https://github.com/Maximus5/ConEmu/issues/885 - https://sudonull.com/post/61039-Speeding-up-the-transfer-of-processes-and-threads-in-Windows-Infopulse-Ukraine-Blog - https://learn.microsoft.com/en-us/windows/win32/api/winternl/nf-winternl-ntquerysysteminformation
1.0
Windows Process.Name() getFromSnapProcess() performance - Currently `Process.Name()` on Windows calls the `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` syscall to get it's ppid and name: https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L321-L322 https://github.com/shirou/gopsutil/blob/400a453e67c4f3b5a3a446e788285cedebaf04f4/process/process_windows.go#L830-L831 It seems `CreateToolhelp32Snapshot(TH32CS_SNAPPROCESS, pid)` will return all processes in the system despite passing the ProcessID parameter. Although returning a snapshot of all processes is quite fast most of the time (<10ms), it can sometimes be slow (>100ms, sometimes even >1000ms). Can we change the flags like below according to the documentation: (not so sure about the exact flags, especially `TH32CS_SNAPMODULE32`) ``` func getFromSnapProcess(pid int32) (int32, int32, string, error) { // https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot // To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero. // Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again, // specifying its process identifier and the TH32CS_SNAPHEAPLIST or TH32_SNAPMODULE value. snap, err := windows.CreateToolhelp32Snapshot(windows.TH32CS_SNAPHEAPLIST|windows.TH32CS_SNAPMODULE|windows.TH32CS_SNAPMODULE32, uint32(pid)) if err != nil { return 0, 0, "", err } defer windows.CloseHandle(snap) var pe32 windows.ProcessEntry32 pe32.Size = uint32(unsafe.Sizeof(pe32)) if err = windows.Process32First(snap, &pe32); err != nil && err != syscall.ERROR_NO_MORE_FILES { return 0, 0, "", err } for { if pe32.ProcessID == uint32(pid) { szexe := windows.UTF16ToString(pe32.ExeFile[:]) return int32(pe32.ParentProcessID), int32(pe32.Threads), szexe, nil } if err = windows.Process32Next(snap, &pe32); err != nil { break } } return 0, 0, "", fmt.Errorf("couldn't find pid: %d", pid) } ``` or switch to `NtQuerySystemInformation`? ## refs https://learn.microsoft.com/en-us/windows/win32/api/tlhelp32/nf-tlhelp32-createtoolhelp32snapshot: - TH32CS_SNAPPROCESS 0x00000002 | Includes all processes in the system in the snapshot. > To enumerate the heap or module states for all processes, specify TH32CS_SNAPALL and set th32ProcessID to zero. > Then, for each additional process in the snapshot, call CreateToolhelp32Snapshot again, specifying its process identifier and the **TH32CS_SNAPHEAPLIST** or **TH32_SNAPMODULE** value. ## see also - https://github.com/Maximus5/ConEmu/issues/885 - https://sudonull.com/post/61039-Speeding-up-the-transfer-of-processes-and-threads-in-Windows-Infopulse-Ukraine-Blog - https://learn.microsoft.com/en-us/windows/win32/api/winternl/nf-winternl-ntquerysysteminformation
process
windows process name getfromsnapprocess performance currently process name on windows calls the snapprocess pid syscall to get it s ppid and name it seems snapprocess pid will return all processes in the system despite passing the processid parameter although returning a snapshot of all processes is quite fast most of the time sometimes even can we change the flags like below according to the documentation not so sure about the exact flags especially func getfromsnapprocess pid string error to enumerate the heap or module states for all processes specify snapall and set to zero then for each additional process in the snapshot call again specifying its process identifier and the snapheaplist or snapmodule value snap err windows windows snapheaplist windows snapmodule windows pid if err nil return err defer windows closehandle snap var windows size unsafe sizeof if err windows snap err nil err syscall error no more files return err for if processid pid szexe windows exefile return parentprocessid threads szexe nil if err windows snap err nil break return fmt errorf couldn t find pid d pid or switch to ntquerysysteminformation refs snapprocess includes all processes in the system in the snapshot to enumerate the heap or module states for all processes specify snapall and set to zero then for each additional process in the snapshot call again specifying its process identifier and the snapheaplist or snapmodule value see also
1
7,209
10,343,374,395
IssuesEvent
2019-09-04 08:49:21
Hurence/logisland
https://api.github.com/repos/Hurence/logisland
closed
Add new processor URLDecode
processor
# Expected behavior and actual behavior. When a field contains a URL, it can be needed to decode special chars encoded in URL (in UT-8 charset for instance) such that it does not appear encoded and mess up further user data processing. The need is then to be able to specify a list of one or more fields subject to URL decoding and a charset (default being UTF-8). # Steps to reproduce the problem. # Specifications like the version of the project, operating system, or hardware.
1.0
Add new processor URLDecode - # Expected behavior and actual behavior. When a field contains a URL, it can be needed to decode special chars encoded in URL (in UT-8 charset for instance) such that it does not appear encoded and mess up further user data processing. The need is then to be able to specify a list of one or more fields subject to URL decoding and a charset (default being UTF-8). # Steps to reproduce the problem. # Specifications like the version of the project, operating system, or hardware.
process
add new processor urldecode expected behavior and actual behavior when a field contains a url it can be needed to decode special chars encoded in url in ut charset for instance such that it does not appear encoded and mess up further user data processing the need is then to be able to specify a list of one or more fields subject to url decoding and a charset default being utf steps to reproduce the problem specifications like the version of the project operating system or hardware
1
21,538
29,837,211,788
IssuesEvent
2023-06-19 00:37:44
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / Belo Horizonte, Minas Gerais, Brazil] Fullstack Developer (.NET/Javascript) na Coodesh
SALVADOR JAVASCRIPT FULL-STACK CSS3 MVC PLENO SQL GIT BOOTSTRAP WEB API ENTITY FRAMEWORK MOBILE REQUISITOS ASP.NET PROCESSOS GITHUB UMA C QUALIDADE MANUTENÇÃO ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A Revelar RH está em busca de Fullstack Developer para compor seu time!</p> <p>Responsabilidades:</p> <ul> <li>Desenvolvimento e acompanhamento de manutenção evolutivas e corretivas em sistemas desenvolvidos em .NET;</li> <li>Propor soluções e desenvolver novas tecnologias para melhoria dos sistemas;</li> <li>Manter a qualidade do código, versionamento, testes, deploy e participação na priorização de funcionalidades.</li> </ul> <p>Horário de segunda à sexta das 8h às 17h.</p> ## Revelar RH: <p>A Revelar RH é uma empresa inovadora que busca alinhar os objetivos das Empresas que têm carência de profissionais qualificados, com os das pessoas que estão em busca de uma oportunidades no mercado de trabalho. Nosso trabalho visa o aprimoramento e racionalização de processos voltados à Gestão de Pessoas, sendo realizado de forma personalizada para cada cliente. Atualmente atendemos várias empresas de tecnologia e softwares house.&nbsp;</p></p> ## Habilidades: - Asp.Net MVC - C# - Arquitetura de Projetos Mobile ## Local: Belo Horizonte, Minas Gerais, Brazil ## Requisitos: - Superior completo ou em curso de Análise e Desenvolvimento de Sistemas, Engenharia da computação ou áreas afins; - Experiência e conhecimento correspondente ao perfil pleno/sênior; - Experiência com HTML5, CSS3, Bootstrap; - Experiência com SQL SERVER; - Experiência com GIT; - Experiência com ASP.NET MVC, WEB API, C# e Entity framework. ## Benefícios: - Plano de saúde; - Plano odontológico; - Seguro de vida; - Cartão de benefícios Swile (alimentação, auxílio internet, desconto em academia). ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (.NET/Javascript) na Revelar RH](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Categoria Full-Stack
1.0
[Hibrido / Belo Horizonte, Minas Gerais, Brazil] Fullstack Developer (.NET/Javascript) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A Revelar RH está em busca de Fullstack Developer para compor seu time!</p> <p>Responsabilidades:</p> <ul> <li>Desenvolvimento e acompanhamento de manutenção evolutivas e corretivas em sistemas desenvolvidos em .NET;</li> <li>Propor soluções e desenvolver novas tecnologias para melhoria dos sistemas;</li> <li>Manter a qualidade do código, versionamento, testes, deploy e participação na priorização de funcionalidades.</li> </ul> <p>Horário de segunda à sexta das 8h às 17h.</p> ## Revelar RH: <p>A Revelar RH é uma empresa inovadora que busca alinhar os objetivos das Empresas que têm carência de profissionais qualificados, com os das pessoas que estão em busca de uma oportunidades no mercado de trabalho. Nosso trabalho visa o aprimoramento e racionalização de processos voltados à Gestão de Pessoas, sendo realizado de forma personalizada para cada cliente. Atualmente atendemos várias empresas de tecnologia e softwares house.&nbsp;</p></p> ## Habilidades: - Asp.Net MVC - C# - Arquitetura de Projetos Mobile ## Local: Belo Horizonte, Minas Gerais, Brazil ## Requisitos: - Superior completo ou em curso de Análise e Desenvolvimento de Sistemas, Engenharia da computação ou áreas afins; - Experiência e conhecimento correspondente ao perfil pleno/sênior; - Experiência com HTML5, CSS3, Bootstrap; - Experiência com SQL SERVER; - Experiência com GIT; - Experiência com ASP.NET MVC, WEB API, C# e Entity framework. ## Benefícios: - Plano de saúde; - Plano odontológico; - Seguro de vida; - Cartão de benefícios Swile (alimentação, auxílio internet, desconto em academia). ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (.NET/Javascript) na Revelar RH](https://coodesh.com/jobs/desenvolvedor-fullstack-112034631?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Categoria Full-Stack
process
fullstack developer net javascript na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a revelar rh está em busca de fullstack developer para compor seu time responsabilidades desenvolvimento e acompanhamento de manutenção evolutivas e corretivas em sistemas desenvolvidos em net propor soluções e desenvolver novas tecnologias para melhoria dos sistemas manter a qualidade do código versionamento testes deploy e participação na priorização de funcionalidades horário de segunda à sexta das às revelar rh a revelar rh é uma empresa inovadora que busca alinhar os objetivos das empresas que têm carência de profissionais qualificados com os das pessoas que estão em busca de uma oportunidades no mercado de trabalho nosso trabalho visa o aprimoramento e racionalização de processos voltados à gestão de pessoas sendo realizado de forma personalizada para cada cliente atualmente atendemos várias empresas de tecnologia e softwares house nbsp habilidades asp net mvc c arquitetura de projetos mobile local belo horizonte minas gerais brazil requisitos superior completo ou em curso de análise e desenvolvimento de sistemas engenharia da computação ou áreas afins experiência e conhecimento correspondente ao perfil pleno sênior experiência com bootstrap experiência com sql server experiência com git experiência com asp net mvc web api c e entity framework benefícios plano de saúde plano odontológico seguro de vida cartão de benefícios swile alimentação auxílio internet desconto em academia como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado categoria full stack
1
8,918
4,351,323,509
IssuesEvent
2016-07-31 19:55:24
VOREStation/VOREStation
https://api.github.com/repos/VOREStation/VOREStation
closed
Never ending heartbeat sounds.
Pri: 2-Moderate Type: Bug Works in latest build
#### Brief description of the issue The sound of a heart beating never ceases. #### What you expected to happen For it to stop. #### What actually happened It never stopped. #### Steps to reproduce 1. Get gurgled. 2. Enjoy the annoying sound. #### Additional info: - **Server Revision**: Server revision: release - 2016-05-23 3bb482ee56a4ade2e02fbb46fa3746728c9a3b3e - **Anything else you may wish to add** (Location if it's a mapping issue, etc)
1.0
Never ending heartbeat sounds. - #### Brief description of the issue The sound of a heart beating never ceases. #### What you expected to happen For it to stop. #### What actually happened It never stopped. #### Steps to reproduce 1. Get gurgled. 2. Enjoy the annoying sound. #### Additional info: - **Server Revision**: Server revision: release - 2016-05-23 3bb482ee56a4ade2e02fbb46fa3746728c9a3b3e - **Anything else you may wish to add** (Location if it's a mapping issue, etc)
non_process
never ending heartbeat sounds brief description of the issue the sound of a heart beating never ceases what you expected to happen for it to stop what actually happened it never stopped steps to reproduce get gurgled enjoy the annoying sound additional info server revision server revision release anything else you may wish to add location if it s a mapping issue etc
0
19,195
25,321,324,410
IssuesEvent
2022-11-18 04:21:07
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Archlinux: "bazel crashed due to an internal error" when building media pipe
more data needed type: support / not a bug (process) team-OSS
### Description of the bug: When attempting to build [mediapipe](https://github.com/google/mediapipe) on Archlinux, Bazel experiences an internal error. Archlinux bazel v5.3.2 java openjdk 19.0.1 2022-10-18 mediapipe v0.8.11 ` [user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion # requires Bazel 5.2.0 by default [user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu Starting local Bazel server and connecting to it... DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_absl' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_benchmark' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'flatbuffers' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'pybind11_bazel' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_googlesource_code_re2' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_protobuf' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_googletest' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_github_gflags_gflags' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'zlib' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'build_bazel_rules_apple' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'build_bazel_rules_swift' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'build_bazel_apple_support' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'xctestrunner' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'pybind11' because it already exists. WARNING: /home/user/Git/mediapipe/mediapipe/framework/BUILD:54:24: in cc_library rule //mediapipe/framework:calculator_cc_proto: target '//mediapipe/framework:calculator_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use. WARNING: /home/user/Git/mediapipe/mediapipe/framework/tool/BUILD:185:24: in cc_library rule //mediapipe/framework/tool:field_data_cc_proto: target '//mediapipe/framework/tool:field_data_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use. INFO: Analyzed target //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu (140 packages loaded, 5537 targets configured). INFO: Found 1 target... [0 / 8] 3 actions, 0 running [Prepa] BazelWorkspaceStatusAction stable-status.txt [Prepa] Writing file mediapipe/examples/desktop/hand_tracking/hand_tracking_cpu-2.params [Prepa] Creating source manifest for //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu FATAL: bazel crashed due to an internal error. Printing stack trace: java.lang.ExceptionInInitializerError at com.google.devtools.build.lib.actions.ParameterFile.writeContent(ParameterFile.java:118) at com.google.devtools.build.lib.actions.ParameterFile.writeParameterFile(ParameterFile.java:111) at com.google.devtools.build.lib.analysis.actions.ParameterFileWriteAction$ParamFileWriter.writeOutputFile(ParameterFileWriteAction.java:170) at com.google.devtools.build.lib.exec.FileWriteStrategy.beginWriteOutputToFile(FileWriteStrategy.java:58) at com.google.devtools.build.lib.analysis.actions.FileWriteActionContext.beginWriteOutputToFile(FileWriteActionContext.java:49) at com.google.devtools.build.lib.analysis.actions.AbstractFileWriteAction.beginExecution(AbstractFileWriteAction.java:66) at com.google.devtools.build.lib.actions.Action.execute(Action.java:133) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031) at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152) at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169) at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590) at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) at java.base/java.lang.Thread.run(Thread.java:1589) Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make java.lang.String(byte[],byte) accessible: module java.base does not "opens java.lang" to unnamed module @7ed7259e at java.base/java.lang.reflect.AccessibleObject.throwInaccessibleObjectException(AccessibleObject.java:387) at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:363) at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:311) at java.base/java.lang.reflect.Constructor.checkCanSetAccessible(Constructor.java:192) at java.base/java.lang.reflect.Constructor.setAccessible(Constructor.java:185) at com.google.devtools.build.lib.unsafe.StringUnsafe.<init>(StringUnsafe.java:75) at com.google.devtools.build.lib.unsafe.StringUnsafe.initInstance(StringUnsafe.java:56) at com.google.devtools.build.lib.unsafe.StringUnsafe.<clinit>(StringUnsafe.java:37) ... 21 more ` Thanks! ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. [user@system mediapipe]$ git clone https://github.com/google/mediapipe.git [user@system mediapipe]$ cd mediapipe [user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion [user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop ### Which operating system are you running Bazel on? Archlinux ### What is the output of `bazel info release`? release 5.3.2 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? No. ### Any other information, logs, or outputs that you want to share? See description.
1.0
Archlinux: "bazel crashed due to an internal error" when building media pipe - ### Description of the bug: When attempting to build [mediapipe](https://github.com/google/mediapipe) on Archlinux, Bazel experiences an internal error. Archlinux bazel v5.3.2 java openjdk 19.0.1 2022-10-18 mediapipe v0.8.11 ` [user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion # requires Bazel 5.2.0 by default [user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu Starting local Bazel server and connecting to it... DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_absl' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_benchmark' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'flatbuffers' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'pybind11_bazel' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_googlesource_code_re2' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_protobuf' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_google_googletest' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'com_github_gflags_gflags' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'zlib' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'build_bazel_rules_apple' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'build_bazel_rules_swift' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'build_bazel_apple_support' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'xctestrunner' because it already exists. DEBUG: /home/user/.cache/bazel/_bazel_user/76aa75a67bc63428e516562f91dd6704/external/org_tensorflow/third_party/repo.bzl:132:14: Warning: skipping import of repository 'pybind11' because it already exists. WARNING: /home/user/Git/mediapipe/mediapipe/framework/BUILD:54:24: in cc_library rule //mediapipe/framework:calculator_cc_proto: target '//mediapipe/framework:calculator_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use. WARNING: /home/user/Git/mediapipe/mediapipe/framework/tool/BUILD:185:24: in cc_library rule //mediapipe/framework/tool:field_data_cc_proto: target '//mediapipe/framework/tool:field_data_cc_proto' depends on deprecated target '@com_google_protobuf//:cc_wkt_protos': Only for backward compatibility. Do not use. INFO: Analyzed target //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu (140 packages loaded, 5537 targets configured). INFO: Found 1 target... [0 / 8] 3 actions, 0 running [Prepa] BazelWorkspaceStatusAction stable-status.txt [Prepa] Writing file mediapipe/examples/desktop/hand_tracking/hand_tracking_cpu-2.params [Prepa] Creating source manifest for //mediapipe/examples/desktop/hand_tracking:hand_tracking_cpu FATAL: bazel crashed due to an internal error. Printing stack trace: java.lang.ExceptionInInitializerError at com.google.devtools.build.lib.actions.ParameterFile.writeContent(ParameterFile.java:118) at com.google.devtools.build.lib.actions.ParameterFile.writeParameterFile(ParameterFile.java:111) at com.google.devtools.build.lib.analysis.actions.ParameterFileWriteAction$ParamFileWriter.writeOutputFile(ParameterFileWriteAction.java:170) at com.google.devtools.build.lib.exec.FileWriteStrategy.beginWriteOutputToFile(FileWriteStrategy.java:58) at com.google.devtools.build.lib.analysis.actions.FileWriteActionContext.beginWriteOutputToFile(FileWriteActionContext.java:49) at com.google.devtools.build.lib.analysis.actions.AbstractFileWriteAction.beginExecution(AbstractFileWriteAction.java:66) at com.google.devtools.build.lib.actions.Action.execute(Action.java:133) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031) at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152) at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169) at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590) at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) at java.base/java.lang.Thread.run(Thread.java:1589) Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make java.lang.String(byte[],byte) accessible: module java.base does not "opens java.lang" to unnamed module @7ed7259e at java.base/java.lang.reflect.AccessibleObject.throwInaccessibleObjectException(AccessibleObject.java:387) at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:363) at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:311) at java.base/java.lang.reflect.Constructor.checkCanSetAccessible(Constructor.java:192) at java.base/java.lang.reflect.Constructor.setAccessible(Constructor.java:185) at com.google.devtools.build.lib.unsafe.StringUnsafe.<init>(StringUnsafe.java:75) at com.google.devtools.build.lib.unsafe.StringUnsafe.initInstance(StringUnsafe.java:56) at com.google.devtools.build.lib.unsafe.StringUnsafe.<clinit>(StringUnsafe.java:37) ... 21 more ` Thanks! ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. [user@system mediapipe]$ git clone https://github.com/google/mediapipe.git [user@system mediapipe]$ cd mediapipe [user@system mediapipe]$ bazel --version | sed 's/bazel //' >.bazelversion [user@system mediapipe]$ bazel build -c opt --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop ### Which operating system are you running Bazel on? Archlinux ### What is the output of `bazel info release`? release 5.3.2 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? No. ### Any other information, logs, or outputs that you want to share? See description.
process
archlinux bazel crashed due to an internal error when building media pipe description of the bug when attempting to build on archlinux bazel experiences an internal error archlinux bazel java openjdk mediapipe bazel version sed s bazel bazelversion requires bazel by default bazel build c opt define mediapipe disable gpu mediapipe examples desktop hand tracking hand tracking cpu starting local bazel server and connecting to it debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google absl because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google benchmark because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository flatbuffers because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository bazel because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com googlesource code because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google protobuf because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com google googletest because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository com github gflags gflags because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository zlib because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository build bazel rules apple because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository build bazel rules swift because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository build bazel apple support because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository xctestrunner because it already exists debug home user cache bazel bazel user external org tensorflow third party repo bzl warning skipping import of repository because it already exists warning home user git mediapipe mediapipe framework build in cc library rule mediapipe framework calculator cc proto target mediapipe framework calculator cc proto depends on deprecated target com google protobuf cc wkt protos only for backward compatibility do not use warning home user git mediapipe mediapipe framework tool build in cc library rule mediapipe framework tool field data cc proto target mediapipe framework tool field data cc proto depends on deprecated target com google protobuf cc wkt protos only for backward compatibility do not use info analyzed target mediapipe examples desktop hand tracking hand tracking cpu packages loaded targets configured info found target actions running bazelworkspacestatusaction stable status txt writing file mediapipe examples desktop hand tracking hand tracking cpu params creating source manifest for mediapipe examples desktop hand tracking hand tracking cpu fatal bazel crashed due to an internal error printing stack trace java lang exceptionininitializererror at com google devtools build lib actions parameterfile writecontent parameterfile java at com google devtools build lib actions parameterfile writeparameterfile parameterfile java at com google devtools build lib analysis actions parameterfilewriteaction paramfilewriter writeoutputfile parameterfilewriteaction java at com google devtools build lib exec filewritestrategy beginwriteoutputtofile filewritestrategy java at com google devtools build lib analysis actions filewriteactioncontext beginwriteoutputtofile filewriteactioncontext java at com google devtools build lib analysis actions abstractfilewriteaction beginexecution abstractfilewriteaction java at com google devtools build lib actions action execute action java at com google devtools build lib skyframe skyframeactionexecutor execute skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner continueaction skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner run skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionstate runstatemachine actionexecutionstate java at com google devtools build lib skyframe actionexecutionstate getresultordependonfuture actionexecutionstate java at com google devtools build lib skyframe skyframeactionexecutor executeaction skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionfunction checkcacheandexecuteifneeded actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction computeinternal actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction compute actionexecutionfunction java at com google devtools build skyframe abstractparallelevaluator evaluate run abstractparallelevaluator java at com google devtools build lib concurrent abstractqueuevisitor wrappedrunnable run abstractqueuevisitor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java lang reflect inaccessibleobjectexception unable to make java lang string byte byte accessible module java base does not opens java lang to unnamed module at java base java lang reflect accessibleobject throwinaccessibleobjectexception accessibleobject java at java base java lang reflect accessibleobject checkcansetaccessible accessibleobject java at java base java lang reflect accessibleobject checkcansetaccessible accessibleobject java at java base java lang reflect constructor checkcansetaccessible constructor java at java base java lang reflect constructor setaccessible constructor java at com google devtools build lib unsafe stringunsafe stringunsafe java at com google devtools build lib unsafe stringunsafe initinstance stringunsafe java at com google devtools build lib unsafe stringunsafe stringunsafe java more thanks what s the simplest easiest way to reproduce this bug please provide a minimal example if possible git clone cd mediapipe bazel version sed s bazel bazelversion bazel build c opt define mediapipe disable gpu mediapipe examples desktop which operating system are you running bazel on archlinux what is the output of bazel info release release if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no any other information logs or outputs that you want to share see description
1
73,001
3,398,568,619
IssuesEvent
2015-12-02 05:01:39
tomreece/pucauto
https://api.github.com/repos/tomreece/pucauto
closed
Make easier to run for Windows users
high priority large
@droogans How do Windows users typically install dependencies that us Mac users would install with pip? I want them to be able to use my `requirements.txt` file not have to manually go grab every dependency I keep adding. I'm about to merge a PR that adds Six to the project. Thoughts?
1.0
Make easier to run for Windows users - @droogans How do Windows users typically install dependencies that us Mac users would install with pip? I want them to be able to use my `requirements.txt` file not have to manually go grab every dependency I keep adding. I'm about to merge a PR that adds Six to the project. Thoughts?
non_process
make easier to run for windows users droogans how do windows users typically install dependencies that us mac users would install with pip i want them to be able to use my requirements txt file not have to manually go grab every dependency i keep adding i m about to merge a pr that adds six to the project thoughts
0
22,106
30,635,987,384
IssuesEvent
2023-07-24 17:48:16
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
roblox-pyc 1.24.102 has 5 GuardDog issues
guarddog silent-process-execution
https://pypi.org/project/roblox-pyc https://inspector.pypi.io/project/roblox-pyc ```{ "dependency": "roblox-pyc", "version": "1.24.102", "result": { "issues": 5, "errors": {}, "results": { "silent-process-execution": [ { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:143", "code": " subprocess.call([\"npm\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:149", "code": " subprocess.call([\"rbxtsc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:188", "code": " subprocess.call([\"wally\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:198", "code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:205", "code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpntaz4g58/roblox-pyc" } }```
1.0
roblox-pyc 1.24.102 has 5 GuardDog issues - https://pypi.org/project/roblox-pyc https://inspector.pypi.io/project/roblox-pyc ```{ "dependency": "roblox-pyc", "version": "1.24.102", "result": { "issues": 5, "errors": {}, "results": { "silent-process-execution": [ { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:143", "code": " subprocess.call([\"npm\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:149", "code": " subprocess.call([\"rbxtsc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:188", "code": " subprocess.call([\"wally\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:198", "code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" }, { "location": "roblox-pyc-1.24.102/robloxpyc/robloxpy.py:205", "code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpntaz4g58/roblox-pyc" } }```
process
roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp roblox pyc
1
238,574
26,135,891,805
IssuesEvent
2022-12-29 12:04:39
MatBenfield/news
https://api.github.com/repos/MatBenfield/news
closed
[SecurityWeek] EarSpy: Spying on Phone Calls via Ear Speaker Vibrations Captured by Accelerometer
SecurityWeek Stale
**As smartphone manufacturers are improving the ear speakers in their devices, it can become easier for malicious actors to leverage a particular side-channel for eavesdropping on a targeted user’s conversations, according to a team of researchers from several universities in the United States.** [read more](https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer) <https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer>
True
[SecurityWeek] EarSpy: Spying on Phone Calls via Ear Speaker Vibrations Captured by Accelerometer - **As smartphone manufacturers are improving the ear speakers in their devices, it can become easier for malicious actors to leverage a particular side-channel for eavesdropping on a targeted user’s conversations, according to a team of researchers from several universities in the United States.** [read more](https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer) <https://www.securityweek.com/earspy-spying-phone-calls-ear-speaker-vibrations-captured-accelerometer>
non_process
earspy spying on phone calls via ear speaker vibrations captured by accelerometer as smartphone manufacturers are improving the ear speakers in their devices it can become easier for malicious actors to leverage a particular side channel for eavesdropping on a targeted user’s conversations according to a team of researchers from several universities in the united states
0