Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
207,876
15,855,835,399
IssuesEvent
2021-04-08 00:51:45
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz
AWS CI CLT Especialista GraphQL NoSQL Python Remoto Rest Stale Testes automatizados
## Descrição da vaga: <p>Nosso propósito aqui no Méliuz é desbloquear relações inteligentes de consumo e para fazermos isso todos os dias contamos com a nossa Grande Família.</p> <p>Estamos entre as melhores empresas para se trabalhar no ranking empresas de Tecnologia e no Ranking Minas Gerais, segundo o Great Place to Work. Isso significa que temos o compromisso de conduzir processos alinhados com os nossos valores, promovendo o bem estar de todos que se relacionam conosco, com foco em objetivos desafiadores. Afinal, queremos alçar vôos muito mais altos.</p> <p>Temos grandes sonhos e desafios e sabemos que para chegarmos cada vez mais longe precisamos aumentar nossa grande família. Especificamente para esta vaga, estamos em busca de pessoas de nível Sênior/Especialista em tecnologias Back-end que nos ajudarão a continuar aprimorando nossa arquitetura voltada a microsserviços e a construir aplicações cada vez mais escaláveis. Temos produtos acessados por milhões de pessoas e muitos projetos para torná-los ainda mais interessantes para nossos usuários e parceiros.</p> <p><strong>😏Já pensou em trabalhar remotamente, de qualquer lugar do Brasil ou de um dos nossos escritórios em Belo Horizonte ou Manaus, e ainda impactar positivamente na vida de milhões de usuários? 🙌</strong></p> <p></p> <p><span style="font-size: 18px;"><strong>RESPONSABILIDADES E ATRIBUIÇÕES</strong></span></p> <ul> <li>Colaborar com nossos times de engenharia, produto e negócio na construção dos nossos produtos;</li> <li>Atuar com protagonismo no desenvolvimento e evolução das nossas tecnologias backend;</li> <li>Projetar soluções e arquiteturas escaláveis para atender grandes volumes de acessos e tráfego de dados;</li> <li>Projetar e implementar a persistência de dados das aplicações e garantir que o acesso seja seguro e eficiente;</li> <li>Arquitetar soluções para problemas complexos e garantir que o código seja eficiente e de fácil manutenção;</li> <li>Atuar como mentor e disseminar conhecimento para outros membros do time.</li> </ul> <p></p> <p><strong>MAIS IMPORTANTE AINDA:</strong></p> <p>👩‍💻Todas as nossas vagas estão disponíveis para pessoas com deficiência!</p> ## Méliuz: <p><span style="color: rgb(30,32,34);background-color: rgb(255,255,255);font-size: 16px;font-family: Poppins, Helvetica, Arial, sans-serif;">O Méliuz é uma empresa que desbloqueia relações inteligentes de consumo. Estamos em uma busca constante de melhorar a forma que nossos usuários fazem suas compras em nossas lojas parceiras, de maneira que todos saiam ganhando. Queremos que nossos parceiros continuem conosco para toda a vida e, para isso, estamos sempre levando inovações até eles. Tudo isso aliado ao nosso espírito empreendedor que nos faz evoluir cada dia mais.</span>&nbsp;</p> </p> ## Habilidades: - AWS - Node.js - CI/CD - NoSQL ## Local: 100% Remoto ## Requisitos: - Conhecimentos avançados em linguagens e tecnologias como Python e Node.js; - Conhecimentos avançados de algoritmos e estruturas de dados; - Experiência trabalhando com banco de dados relacionais e NoSQL; - Experiência sólida desenvolvendo APIs com REST, gRPC ou GraphQL; - Experiência utilizando cache (elasticache) e serviços de fila(SQS/SNS); - Arquitetura orientada a eventos e microsserviços; - Experiência com testes automatizados e ambientes de CI/CD; - Experiência implementando arquiteturas escaláveis em ambiente cloud, principalmente AWS. ## Benefícios: - 🚈 Vale transporte; - 🥗 Vale refeição ou alimentação; - 🤩Plano de saúde; - 😁 Plano odontológico; - ⏰ Horário flexível; - 💰 Premiação por atingimento de metas coletivas; - 🏊‍♀ Convênio com Sesc; - 📚 Biblioteca interna; - 🤑 Cashback online em dobro; - 🤰 Licença Maternidade/Paternidade estendida; - 💻 Ajuda de custo para o teletrabalho; - 👶 Auxílio Creche; - ✝ Auxílio luto. ## Como se candidatar: Candidate-se em: [Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz](https://coodesh.com/vagas/pessoa-desenvolvedora-backend-seniorespecialista-vaga-remota-002652?origin=github&modal=open) ## Tempo médio de feedbacks: Costumamos enviar feedbacks em até 48 horas após cada etapa do processo. E-mail para contato em caso de não haver resposta: [tamara.carvalho@meliuz.com.br](mailto:tamara.carvalho@meliuz.com.br) ## Labels #### Alocação Remoto #### Regime CLT #### Nível Sênior
1.0
[Remoto] Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz - ## Descrição da vaga: <p>Nosso propósito aqui no Méliuz é desbloquear relações inteligentes de consumo e para fazermos isso todos os dias contamos com a nossa Grande Família.</p> <p>Estamos entre as melhores empresas para se trabalhar no ranking empresas de Tecnologia e no Ranking Minas Gerais, segundo o Great Place to Work. Isso significa que temos o compromisso de conduzir processos alinhados com os nossos valores, promovendo o bem estar de todos que se relacionam conosco, com foco em objetivos desafiadores. Afinal, queremos alçar vôos muito mais altos.</p> <p>Temos grandes sonhos e desafios e sabemos que para chegarmos cada vez mais longe precisamos aumentar nossa grande família. Especificamente para esta vaga, estamos em busca de pessoas de nível Sênior/Especialista em tecnologias Back-end que nos ajudarão a continuar aprimorando nossa arquitetura voltada a microsserviços e a construir aplicações cada vez mais escaláveis. Temos produtos acessados por milhões de pessoas e muitos projetos para torná-los ainda mais interessantes para nossos usuários e parceiros.</p> <p><strong>😏Já pensou em trabalhar remotamente, de qualquer lugar do Brasil ou de um dos nossos escritórios em Belo Horizonte ou Manaus, e ainda impactar positivamente na vida de milhões de usuários? 🙌</strong></p> <p></p> <p><span style="font-size: 18px;"><strong>RESPONSABILIDADES E ATRIBUIÇÕES</strong></span></p> <ul> <li>Colaborar com nossos times de engenharia, produto e negócio na construção dos nossos produtos;</li> <li>Atuar com protagonismo no desenvolvimento e evolução das nossas tecnologias backend;</li> <li>Projetar soluções e arquiteturas escaláveis para atender grandes volumes de acessos e tráfego de dados;</li> <li>Projetar e implementar a persistência de dados das aplicações e garantir que o acesso seja seguro e eficiente;</li> <li>Arquitetar soluções para problemas complexos e garantir que o código seja eficiente e de fácil manutenção;</li> <li>Atuar como mentor e disseminar conhecimento para outros membros do time.</li> </ul> <p></p> <p><strong>MAIS IMPORTANTE AINDA:</strong></p> <p>👩‍💻Todas as nossas vagas estão disponíveis para pessoas com deficiência!</p> ## Méliuz: <p><span style="color: rgb(30,32,34);background-color: rgb(255,255,255);font-size: 16px;font-family: Poppins, Helvetica, Arial, sans-serif;">O Méliuz é uma empresa que desbloqueia relações inteligentes de consumo. Estamos em uma busca constante de melhorar a forma que nossos usuários fazem suas compras em nossas lojas parceiras, de maneira que todos saiam ganhando. Queremos que nossos parceiros continuem conosco para toda a vida e, para isso, estamos sempre levando inovações até eles. Tudo isso aliado ao nosso espírito empreendedor que nos faz evoluir cada dia mais.</span>&nbsp;</p> </p> ## Habilidades: - AWS - Node.js - CI/CD - NoSQL ## Local: 100% Remoto ## Requisitos: - Conhecimentos avançados em linguagens e tecnologias como Python e Node.js; - Conhecimentos avançados de algoritmos e estruturas de dados; - Experiência trabalhando com banco de dados relacionais e NoSQL; - Experiência sólida desenvolvendo APIs com REST, gRPC ou GraphQL; - Experiência utilizando cache (elasticache) e serviços de fila(SQS/SNS); - Arquitetura orientada a eventos e microsserviços; - Experiência com testes automatizados e ambientes de CI/CD; - Experiência implementando arquiteturas escaláveis em ambiente cloud, principalmente AWS. ## Benefícios: - 🚈 Vale transporte; - 🥗 Vale refeição ou alimentação; - 🤩Plano de saúde; - 😁 Plano odontológico; - ⏰ Horário flexível; - 💰 Premiação por atingimento de metas coletivas; - 🏊‍♀ Convênio com Sesc; - 📚 Biblioteca interna; - 🤑 Cashback online em dobro; - 🤰 Licença Maternidade/Paternidade estendida; - 💻 Ajuda de custo para o teletrabalho; - 👶 Auxílio Creche; - ✝ Auxílio luto. ## Como se candidatar: Candidate-se em: [Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz](https://coodesh.com/vagas/pessoa-desenvolvedora-backend-seniorespecialista-vaga-remota-002652?origin=github&modal=open) ## Tempo médio de feedbacks: Costumamos enviar feedbacks em até 48 horas após cada etapa do processo. E-mail para contato em caso de não haver resposta: [tamara.carvalho@meliuz.com.br](mailto:tamara.carvalho@meliuz.com.br) ## Labels #### Alocação Remoto #### Regime CLT #### Nível Sênior
test
pessoa desenvolvedora back end sênior especialista vaga remota na méliuz descrição da vaga nosso propósito aqui no méliuz é desbloquear relações inteligentes de consumo e para fazermos isso todos os dias contamos com a nossa grande família estamos entre as melhores empresas para se trabalhar no ranking empresas de tecnologia e no ranking minas gerais segundo o great place to work isso significa que temos o compromisso de conduzir processos alinhados com os nossos valores promovendo o bem estar de todos que se relacionam conosco com foco em objetivos desafiadores afinal queremos alçar vôos muito mais altos temos grandes sonhos e desafios e sabemos que para chegarmos cada vez mais longe precisamos aumentar nossa grande família especificamente para esta vaga estamos em busca de pessoas de nível sênior especialista em tecnologias back end que nos ajudarão a continuar aprimorando nossa arquitetura voltada a microsserviços e a construir aplicações cada vez mais escaláveis temos produtos acessados por milhões de pessoas e muitos projetos para torná los ainda mais interessantes para nossos usuários e parceiros 😏já pensou em trabalhar remotamente de qualquer lugar do brasil ou de um dos nossos escritórios em belo horizonte ou manaus e ainda impactar positivamente na vida de milhões de usuários 🙌 responsabilidades e atribuições colaborar com nossos times de engenharia produto e negócio na construção dos nossos produtos atuar com protagonismo no desenvolvimento e evolução das nossas tecnologias backend projetar soluções e arquiteturas escaláveis para atender grandes volumes de acessos e tráfego de dados projetar e implementar a persistência de dados das aplicações e garantir que o acesso seja seguro e eficiente arquitetar soluções para problemas complexos e garantir que o código seja eficiente e de fácil manutenção atuar como mentor e disseminar conhecimento para outros membros do time mais importante ainda 👩‍💻todas as nossas vagas estão disponíveis para pessoas com deficiência méliuz o méliuz é uma empresa que desbloqueia relações inteligentes de consumo estamos em uma busca constante de melhorar a forma que nossos usuários fazem suas compras em nossas lojas parceiras de maneira que todos saiam ganhando queremos que nossos parceiros continuem conosco para toda a vida e para isso estamos sempre levando inovações até eles tudo isso aliado ao nosso espírito empreendedor que nos faz evoluir cada dia mais nbsp habilidades aws node js ci cd nosql local remoto requisitos conhecimentos avançados em linguagens e tecnologias como python e node js conhecimentos avançados de algoritmos e estruturas de dados experiência trabalhando com banco de dados relacionais e nosql experiência sólida desenvolvendo apis com rest grpc ou graphql experiência utilizando cache elasticache e serviços de fila sqs sns arquitetura orientada a eventos e microsserviços experiência com testes automatizados e ambientes de ci cd experiência implementando arquiteturas escaláveis em ambiente cloud principalmente aws benefícios 🚈 vale transporte 🥗 vale refeição ou alimentação 🤩plano de saúde 😁 plano odontológico ⏰ horário flexível 💰 premiação por atingimento de metas coletivas 🏊‍♀ convênio com sesc 📚 biblioteca interna 🤑 cashback online em dobro 🤰 licença maternidade paternidade estendida 💻 ajuda de custo para o teletrabalho 👶 auxílio creche ✝ auxílio luto como se candidatar candidate se em tempo médio de feedbacks costumamos enviar feedbacks em até horas após cada etapa do processo e mail para contato em caso de não haver resposta mailto tamara carvalho meliuz com br labels alocação remoto regime clt nível sênior
1
158,691
12,422,628,872
IssuesEvent
2020-05-23 23:29:11
sqlalchemy/sqlalchemy
https://api.github.com/repos/sqlalchemy/sqlalchemy
closed
Update profiles to python 3.8
tests
The ci now uses python 3.8 but the profiles are still for python 3.7.
1.0
Update profiles to python 3.8 - The ci now uses python 3.8 but the profiles are still for python 3.7.
test
update profiles to python the ci now uses python but the profiles are still for python
1
419,862
12,229,305,440
IssuesEvent
2020-05-03 23:31:46
roed314/seminars
https://api.github.com/repos/roed314/seminars
opened
Declutter view talk page
low priority user interface
On each "View talk" page: 1) The "Livestream access" box does not deserve to have its own line. It should be moved to the right of the text "Mon May 4, 07:30-08:30 (starts in 12 hours)". 2) Maybe we could remove the word "Abstract" since in essentially all cases, it is obvious that the abstract is the abstract. 3) I think we could simplify Talk content * paper * slides * video to (paper | slides | video) which simplifies further to (slides | video) (slides) or nothing at all, depending on how many items are present. Or if we really want to keep the words "Talk content", at least put everything on one line: Talk content: (paper | slides | video)
1.0
Declutter view talk page - On each "View talk" page: 1) The "Livestream access" box does not deserve to have its own line. It should be moved to the right of the text "Mon May 4, 07:30-08:30 (starts in 12 hours)". 2) Maybe we could remove the word "Abstract" since in essentially all cases, it is obvious that the abstract is the abstract. 3) I think we could simplify Talk content * paper * slides * video to (paper | slides | video) which simplifies further to (slides | video) (slides) or nothing at all, depending on how many items are present. Or if we really want to keep the words "Talk content", at least put everything on one line: Talk content: (paper | slides | video)
non_test
declutter view talk page on each view talk page the livestream access box does not deserve to have its own line it should be moved to the right of the text mon may starts in hours maybe we could remove the word abstract since in essentially all cases it is obvious that the abstract is the abstract i think we could simplify talk content paper slides video to paper slides video which simplifies further to slides video slides or nothing at all depending on how many items are present or if we really want to keep the words talk content at least put everything on one line talk content paper slides video
0
60,000
8,393,874,120
IssuesEvent
2018-10-09 21:57:10
mfem/mfem
https://api.github.com/repos/mfem/mfem
closed
Nedelec bases on prisms and tetrahedra
documentation enhancement fem
Hello, I need to have a discussion about high order Nedelec bases on prisms and tetrahedra. Mainly so that I'm sure I have a reasonable understanding of the requirements before I make another attempt at implementing ND and RT bases for prisms. The ND basis functions associated with edges have a clearly defined orientation which is directed along the edge. Interior basis functions can be defined however we choose. Basis functions on quadrilateral faces can also borrow their orientation from the bounding edges. However, the basis functions associated with triangular faces must align with two of the three bounding edges and its important that neighboring elements choose the same pair of edges. The tetrahedral implementation in MFEM reorders the vertices of each tetrahedron so that the two lowest vertex indices are first. The ND basis functions on triangular faces are then defined so that they align with the two edges which meet at the vertex with the lowest index on each face. This effectively determines the basis function orientations for every face in a tetrahedral mesh. Moreover, face 0 of every tetrahedron in the mesh will choose the same pair of local edges to define its basis functions, and likewise for faces 1, 2, and 3. I don't believe there's a way to define prisms so that they posses this same convenient behavior. If we choose the basis functions on the base of the prism in a certain way there are still three possible choices for the orientation of the basis functions on the top of the prism. In effect this produces three types of prisms in a general mesh. This is unlike what we see for tetrahedra or hexahedra in MFEM. At the very least I expect I'll need to introduce more flexibility into how we handle DOF permutations when working with prisms. What I don't know is just how well I can contain these changes. Any thoughts on potential solutions or pitfalls would be appreciated. Thanks! Mark
1.0
Nedelec bases on prisms and tetrahedra - Hello, I need to have a discussion about high order Nedelec bases on prisms and tetrahedra. Mainly so that I'm sure I have a reasonable understanding of the requirements before I make another attempt at implementing ND and RT bases for prisms. The ND basis functions associated with edges have a clearly defined orientation which is directed along the edge. Interior basis functions can be defined however we choose. Basis functions on quadrilateral faces can also borrow their orientation from the bounding edges. However, the basis functions associated with triangular faces must align with two of the three bounding edges and its important that neighboring elements choose the same pair of edges. The tetrahedral implementation in MFEM reorders the vertices of each tetrahedron so that the two lowest vertex indices are first. The ND basis functions on triangular faces are then defined so that they align with the two edges which meet at the vertex with the lowest index on each face. This effectively determines the basis function orientations for every face in a tetrahedral mesh. Moreover, face 0 of every tetrahedron in the mesh will choose the same pair of local edges to define its basis functions, and likewise for faces 1, 2, and 3. I don't believe there's a way to define prisms so that they posses this same convenient behavior. If we choose the basis functions on the base of the prism in a certain way there are still three possible choices for the orientation of the basis functions on the top of the prism. In effect this produces three types of prisms in a general mesh. This is unlike what we see for tetrahedra or hexahedra in MFEM. At the very least I expect I'll need to introduce more flexibility into how we handle DOF permutations when working with prisms. What I don't know is just how well I can contain these changes. Any thoughts on potential solutions or pitfalls would be appreciated. Thanks! Mark
non_test
nedelec bases on prisms and tetrahedra hello i need to have a discussion about high order nedelec bases on prisms and tetrahedra mainly so that i m sure i have a reasonable understanding of the requirements before i make another attempt at implementing nd and rt bases for prisms the nd basis functions associated with edges have a clearly defined orientation which is directed along the edge interior basis functions can be defined however we choose basis functions on quadrilateral faces can also borrow their orientation from the bounding edges however the basis functions associated with triangular faces must align with two of the three bounding edges and its important that neighboring elements choose the same pair of edges the tetrahedral implementation in mfem reorders the vertices of each tetrahedron so that the two lowest vertex indices are first the nd basis functions on triangular faces are then defined so that they align with the two edges which meet at the vertex with the lowest index on each face this effectively determines the basis function orientations for every face in a tetrahedral mesh moreover face of every tetrahedron in the mesh will choose the same pair of local edges to define its basis functions and likewise for faces and i don t believe there s a way to define prisms so that they posses this same convenient behavior if we choose the basis functions on the base of the prism in a certain way there are still three possible choices for the orientation of the basis functions on the top of the prism in effect this produces three types of prisms in a general mesh this is unlike what we see for tetrahedra or hexahedra in mfem at the very least i expect i ll need to introduce more flexibility into how we handle dof permutations when working with prisms what i don t know is just how well i can contain these changes any thoughts on potential solutions or pitfalls would be appreciated thanks mark
0
149,441
13,281,333,405
IssuesEvent
2020-08-23 16:50:54
genedan/TmVal
https://api.github.com/repos/genedan/TmVal
closed
Rework simple interest examples in docs
documentation
Since SimpleAmt has been deprecated, we need new examples
1.0
Rework simple interest examples in docs - Since SimpleAmt has been deprecated, we need new examples
non_test
rework simple interest examples in docs since simpleamt has been deprecated we need new examples
0
216,219
16,748,249,928
IssuesEvent
2021-06-11 18:34:10
firebase/firebase-cpp-sdk
https://api.github.com/repos/firebase/firebase-cpp-sdk
closed
Nightly Integration Testing Report
nightly-testing tests: failed
### ❌&nbsp; Integration test FAILED Requested by @sunmou99 on commit f7031c96a736f4bf4a56c9a164f261f9ec0d80e4 Last updated: Fri Jun 11 02:42 PDT 2021 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/927716699)** | Failures | Configs | |----------|---------| | auth | <details><summary>(5 items)</summary>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/>[TEST] [FAILURE] [Android] [macos, windows] [emulator_min]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Android] [ubuntu] [emulator_target]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details>[TEST] [FAILURE] [MacOS] [All ssl_lib]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Windows] [openssl]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details></details> | | database | [TEST] [ERROR] [iOS] [macos] [All Simulators]<br/> | | dynamic_links | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest]<br/> | | firestore | <details><summary>(4 items)</summary>[TEST] [ERROR] [Android] [All os] [emulator_min, android_min]<br/>[TEST] [ERROR] [Android] [macos] [emulator_target]<br/>[TEST] [ERROR] [Android] [ubuntu] [android_target]<br/>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/></details> | | functions | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(7 failed tests)</summary>&nbsp;&nbsp;TestErrorHandling<br/>&nbsp;&nbsp;TestFunction<br/>&nbsp;&nbsp;TestFunctionWithAuthToken<br/>&nbsp;&nbsp;TestFunctionWithData<br/>&nbsp;&nbsp;TestFunctionWithNull<br/>&nbsp;&nbsp;TestFunctionWithScalar<br/>&nbsp;&nbsp;TestSignIn</details> | | installations | [TEST] [ERROR] [iOS] [macos] [simulator_target]<br/>[TEST] [FAILURE] [Android] [windows] [android_min]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestGettingTokenTwiceMatches</details>[TEST] [FAILURE] [iOS] [macos] [simulator_latest, simulator_min]<details><summary>(8 failed tests)</summary>&nbsp;&nbsp;TestCanGetId<br/>&nbsp;&nbsp;TestCanGetIdAndTokenTogether<br/>&nbsp;&nbsp;TestCanGetToken<br/>&nbsp;&nbsp;TestDeleteGivesNewIdNextTime<br/>&nbsp;&nbsp;TestDeleteGivesNewTokenNextTime<br/>&nbsp;&nbsp;TestGetTokenForceRefresh<br/>&nbsp;&nbsp;TestGettingIdTwiceMatches<br/>&nbsp;&nbsp;TestGettingTokenTwiceMatches</details> | | messaging | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest, android_target, android_min]<br/> | | remote_config | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(2 failed tests)</summary>&nbsp;&nbsp;TestFetchV2<br/>&nbsp;&nbsp;TestGetAll</details> | | storage | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(12 failed tests)</summary>&nbsp;&nbsp;TestCreateWorkingFolder<br/>&nbsp;&nbsp;TestDeleteFile<br/>&nbsp;&nbsp;TestDownloadUrl<br/>&nbsp;&nbsp;TestInvalidatingReferencesWhenDeletingApp<br/>&nbsp;&nbsp;TestInvalidatingReferencesWhenDeletingStorage<br/>&nbsp;&nbsp;TestLargeFileCancelUpload<br/>&nbsp;&nbsp;TestLargeFilePauseResumeAndDownloadCancel<br/>&nbsp;&nbsp;TestPutFileAndGetFile<br/>&nbsp;&nbsp;TestSignIn<br/>&nbsp;&nbsp;TestStorageUrl<br/>&nbsp;&nbsp;TestWriteAndReadByteBuffer<br/>&nbsp;&nbsp;TestWriteAndReadFileWithCustomMetadata</details> | <hidden value="integration-test-status-comment"></hidden>
2.0
Nightly Integration Testing Report - ### ❌&nbsp; Integration test FAILED Requested by @sunmou99 on commit f7031c96a736f4bf4a56c9a164f261f9ec0d80e4 Last updated: Fri Jun 11 02:42 PDT 2021 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/927716699)** | Failures | Configs | |----------|---------| | auth | <details><summary>(5 items)</summary>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/>[TEST] [FAILURE] [Android] [macos, windows] [emulator_min]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Android] [ubuntu] [emulator_target]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details>[TEST] [FAILURE] [MacOS] [All ssl_lib]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Windows] [openssl]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestSendPasswordResetEmail</details></details> | | database | [TEST] [ERROR] [iOS] [macos] [All Simulators]<br/> | | dynamic_links | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest]<br/> | | firestore | <details><summary>(4 items)</summary>[TEST] [ERROR] [Android] [All os] [emulator_min, android_min]<br/>[TEST] [ERROR] [Android] [macos] [emulator_target]<br/>[TEST] [ERROR] [Android] [ubuntu] [android_target]<br/>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/></details> | | functions | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(7 failed tests)</summary>&nbsp;&nbsp;TestErrorHandling<br/>&nbsp;&nbsp;TestFunction<br/>&nbsp;&nbsp;TestFunctionWithAuthToken<br/>&nbsp;&nbsp;TestFunctionWithData<br/>&nbsp;&nbsp;TestFunctionWithNull<br/>&nbsp;&nbsp;TestFunctionWithScalar<br/>&nbsp;&nbsp;TestSignIn</details> | | installations | [TEST] [ERROR] [iOS] [macos] [simulator_target]<br/>[TEST] [FAILURE] [Android] [windows] [android_min]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;TestGettingTokenTwiceMatches</details>[TEST] [FAILURE] [iOS] [macos] [simulator_latest, simulator_min]<details><summary>(8 failed tests)</summary>&nbsp;&nbsp;TestCanGetId<br/>&nbsp;&nbsp;TestCanGetIdAndTokenTogether<br/>&nbsp;&nbsp;TestCanGetToken<br/>&nbsp;&nbsp;TestDeleteGivesNewIdNextTime<br/>&nbsp;&nbsp;TestDeleteGivesNewTokenNextTime<br/>&nbsp;&nbsp;TestGetTokenForceRefresh<br/>&nbsp;&nbsp;TestGettingIdTwiceMatches<br/>&nbsp;&nbsp;TestGettingTokenTwiceMatches</details> | | messaging | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest, android_target, android_min]<br/> | | remote_config | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(2 failed tests)</summary>&nbsp;&nbsp;TestFetchV2<br/>&nbsp;&nbsp;TestGetAll</details> | | storage | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(12 failed tests)</summary>&nbsp;&nbsp;TestCreateWorkingFolder<br/>&nbsp;&nbsp;TestDeleteFile<br/>&nbsp;&nbsp;TestDownloadUrl<br/>&nbsp;&nbsp;TestInvalidatingReferencesWhenDeletingApp<br/>&nbsp;&nbsp;TestInvalidatingReferencesWhenDeletingStorage<br/>&nbsp;&nbsp;TestLargeFileCancelUpload<br/>&nbsp;&nbsp;TestLargeFilePauseResumeAndDownloadCancel<br/>&nbsp;&nbsp;TestPutFileAndGetFile<br/>&nbsp;&nbsp;TestSignIn<br/>&nbsp;&nbsp;TestStorageUrl<br/>&nbsp;&nbsp;TestWriteAndReadByteBuffer<br/>&nbsp;&nbsp;TestWriteAndReadFileWithCustomMetadata</details> | <hidden value="integration-test-status-comment"></hidden>
test
nightly integration testing report ❌ nbsp integration test failed requested by on commit last updated fri jun pdt failures configs auth items failed tests nbsp nbsp testsendpasswordresetemail failed tests nbsp nbsp testsendpasswordresetemail failed tests nbsp nbsp testsendpasswordresetemail failed tests nbsp nbsp testsendpasswordresetemail database dynamic links firestore items functions failed tests nbsp nbsp testerrorhandling nbsp nbsp testfunction nbsp nbsp testfunctionwithauthtoken nbsp nbsp testfunctionwithdata nbsp nbsp testfunctionwithnull nbsp nbsp testfunctionwithscalar nbsp nbsp testsignin installations failed tests nbsp nbsp testgettingtokentwicematches failed tests nbsp nbsp testcangetid nbsp nbsp testcangetidandtokentogether nbsp nbsp testcangettoken nbsp nbsp testdeletegivesnewidnexttime nbsp nbsp testdeletegivesnewtokennexttime nbsp nbsp testgettokenforcerefresh nbsp nbsp testgettingidtwicematches nbsp nbsp testgettingtokentwicematches messaging remote config failed tests nbsp nbsp nbsp nbsp testgetall storage failed tests nbsp nbsp testcreateworkingfolder nbsp nbsp testdeletefile nbsp nbsp testdownloadurl nbsp nbsp testinvalidatingreferenceswhendeletingapp nbsp nbsp testinvalidatingreferenceswhendeletingstorage nbsp nbsp testlargefilecancelupload nbsp nbsp testlargefilepauseresumeanddownloadcancel nbsp nbsp testputfileandgetfile nbsp nbsp testsignin nbsp nbsp teststorageurl nbsp nbsp testwriteandreadbytebuffer nbsp nbsp testwriteandreadfilewithcustommetadata
1
201,480
15,209,044,840
IssuesEvent
2021-02-17 04:17:37
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] SimpleSecurityNetty4ServerTransportTests.testThreadContext
:Distributed/Network >test-failure Team:Distributed
**Build scan**: https://gradle-enterprise.elastic.co/s/xxazk5c2csshk **Repro line**: ``` ./gradlew ':x-pack:plugin:security:test' --tests "org.elasticsearch.xpack.security.transport.netty4.SimpleSecurityNetty4ServerTransportTests.testThreadContext" -Dtests.seed=3DD775158F20CC58 -Dtests.security.manager=true -Dtests.locale=tr -Dtests.timezone=Etc/GMT-0 -Druntime.java=11 ``` **Reproduces locally?**: No **Applicable branches**: master **Failure history**: **Failure excerpt**: ``` java.lang.AssertionError: this header is only visible in the handler context expected null, but was:<booooom> at __randomizedtesting.SeedInfo.seed([3DD775158F20CC58:ABD4401512CF7604]:0) at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotNull(Assert.java:755) at org.junit.Assert.assertNull(Assert.java:737) at org.elasticsearch.transport.AbstractSimpleTransportTestCase.testThreadContext(AbstractSimpleTransportTestCase.java:383) ```
1.0
[CI] SimpleSecurityNetty4ServerTransportTests.testThreadContext - **Build scan**: https://gradle-enterprise.elastic.co/s/xxazk5c2csshk **Repro line**: ``` ./gradlew ':x-pack:plugin:security:test' --tests "org.elasticsearch.xpack.security.transport.netty4.SimpleSecurityNetty4ServerTransportTests.testThreadContext" -Dtests.seed=3DD775158F20CC58 -Dtests.security.manager=true -Dtests.locale=tr -Dtests.timezone=Etc/GMT-0 -Druntime.java=11 ``` **Reproduces locally?**: No **Applicable branches**: master **Failure history**: **Failure excerpt**: ``` java.lang.AssertionError: this header is only visible in the handler context expected null, but was:<booooom> at __randomizedtesting.SeedInfo.seed([3DD775158F20CC58:ABD4401512CF7604]:0) at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotNull(Assert.java:755) at org.junit.Assert.assertNull(Assert.java:737) at org.elasticsearch.transport.AbstractSimpleTransportTestCase.testThreadContext(AbstractSimpleTransportTestCase.java:383) ```
test
testthreadcontext build scan repro line gradlew x pack plugin security test tests org elasticsearch xpack security transport testthreadcontext dtests seed dtests security manager true dtests locale tr dtests timezone etc gmt druntime java reproduces locally no applicable branches master failure history failure excerpt java lang assertionerror this header is only visible in the handler context expected null but was at  randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert failnotnull assert java at org junit assert assertnull assert java at org elasticsearch transport abstractsimpletransporttestcase testthreadcontext abstractsimpletransporttestcase java
1
127,717
10,479,773,991
IssuesEvent
2019-09-24 05:35:59
linewalks/MDwalks-UI
https://api.github.com/repos/linewalks/MDwalks-UI
closed
chartUtility Test 만들기
DONE test
**개요 및 요청사항** - [x] strIdConvert - [x] tableHeaderConvert - [x] renderSVG - [x] generateGroup - [x] getStartAndEndTime - [x] circleDataFilter - [x] rectDataFilter - [x] labelList **결과물** **완료 기준** - [ ] PR Review @
1.0
chartUtility Test 만들기 - **개요 및 요청사항** - [x] strIdConvert - [x] tableHeaderConvert - [x] renderSVG - [x] generateGroup - [x] getStartAndEndTime - [x] circleDataFilter - [x] rectDataFilter - [x] labelList **결과물** **완료 기준** - [ ] PR Review @
test
chartutility test 만들기 개요 및 요청사항 stridconvert tableheaderconvert rendersvg generategroup getstartandendtime circledatafilter rectdatafilter labellist 결과물 완료 기준 pr review
1
95,649
8,569,239,211
IssuesEvent
2018-11-11 08:18:41
IMA-WorldHealth/bhima-2.X
https://api.github.com/repos/IMA-WorldHealth/bhima-2.X
closed
barcodes.reverseLookup() is broken
Bug needs tests
Looks like we forgot to remove the .done() calls in the controllers/reports.js. Since we use async/await and `.done()` is nonstandard, the two are not compatible. We should makes sure we test our barcodes via integration testing.
1.0
barcodes.reverseLookup() is broken - Looks like we forgot to remove the .done() calls in the controllers/reports.js. Since we use async/await and `.done()` is nonstandard, the two are not compatible. We should makes sure we test our barcodes via integration testing.
test
barcodes reverselookup is broken looks like we forgot to remove the done calls in the controllers reports js since we use async await and done is nonstandard the two are not compatible we should makes sure we test our barcodes via integration testing
1
112,266
4,514,861,102
IssuesEvent
2016-09-05 02:43:07
pombase/canto
https://api.github.com/repos/pombase/canto
closed
Facilitating checking sessions in approval
admin high priority next quick user interface
We want the tability to "tick off" annotations made by the community curator where we agree with their annotations. When they make a lot of annotations it can get a bit confusing otherwise. One suggestion: -highlight all of the annotation (turn it red or something) and then the ability to hit an admin only check box when we check the annotation off.
1.0
Facilitating checking sessions in approval - We want the tability to "tick off" annotations made by the community curator where we agree with their annotations. When they make a lot of annotations it can get a bit confusing otherwise. One suggestion: -highlight all of the annotation (turn it red or something) and then the ability to hit an admin only check box when we check the annotation off.
non_test
facilitating checking sessions in approval we want the tability to tick off annotations made by the community curator where we agree with their annotations when they make a lot of annotations it can get a bit confusing otherwise one suggestion highlight all of the annotation turn it red or something and then the ability to hit an admin only check box when we check the annotation off
0
16,479
2,615,116,951
IssuesEvent
2015-03-01 05:42:09
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
calendar-v2-jsonc-oauth-sample
auto-migrated Priority-Low Type-Sample
``` Which API and version (e.g. Google Calendar Data API version 2)? Google Calendar Data API version 2 What format (e.g. JSON, Atom)? JSON-C What Authentation (e.g. OAuth, OAuth 2, Android, ClientLogin)? OAuth Java environment (e.g. Java 6, Android 2.2, App Engine 1.3.7)? Java 6 External references, such as API reference guide? http://code.google.com/apis/calendar/data/2.0/developers_guide_protocol.html Please provide any additional information below. ``` Original issue reported on code.google.com by `yan...@google.com` on 7 Oct 2010 at 3:07
1.0
calendar-v2-jsonc-oauth-sample - ``` Which API and version (e.g. Google Calendar Data API version 2)? Google Calendar Data API version 2 What format (e.g. JSON, Atom)? JSON-C What Authentation (e.g. OAuth, OAuth 2, Android, ClientLogin)? OAuth Java environment (e.g. Java 6, Android 2.2, App Engine 1.3.7)? Java 6 External references, such as API reference guide? http://code.google.com/apis/calendar/data/2.0/developers_guide_protocol.html Please provide any additional information below. ``` Original issue reported on code.google.com by `yan...@google.com` on 7 Oct 2010 at 3:07
non_test
calendar jsonc oauth sample which api and version e g google calendar data api version google calendar data api version what format e g json atom json c what authentation e g oauth oauth android clientlogin oauth java environment e g java android app engine java external references such as api reference guide please provide any additional information below original issue reported on code google com by yan google com on oct at
0
31,070
11,867,197,766
IssuesEvent
2020-03-26 06:17:02
diecutfacemasks/diecutfacemasks.github.io
https://api.github.com/repos/diecutfacemasks/diecutfacemasks.github.io
closed
Fix TLS for diecutfacemasks.org
bug security ux
https://diecutfacemasks.org creates a certificate warning, this needs to be fixed to ensure user friendly and secure access to the web site.
True
Fix TLS for diecutfacemasks.org - https://diecutfacemasks.org creates a certificate warning, this needs to be fixed to ensure user friendly and secure access to the web site.
non_test
fix tls for diecutfacemasks org creates a certificate warning this needs to be fixed to ensure user friendly and secure access to the web site
0
6,327
3,011,040,529
IssuesEvent
2015-07-28 15:57:07
softlayer/sl-ember-components
https://api.github.com/repos/softlayer/sl-ember-components
closed
Confirm whether sl-panel has a dependency on Bootstrap-Datepicker
2 - Working documentation research
The `#/browsers` page in the demo app indicates as such but I'm thinking this is not accurate. If it is not update the compatibility table. <!--- @huboard:{"milestone_order":364.9999999925494} -->
1.0
Confirm whether sl-panel has a dependency on Bootstrap-Datepicker - The `#/browsers` page in the demo app indicates as such but I'm thinking this is not accurate. If it is not update the compatibility table. <!--- @huboard:{"milestone_order":364.9999999925494} -->
non_test
confirm whether sl panel has a dependency on bootstrap datepicker the browsers page in the demo app indicates as such but i m thinking this is not accurate if it is not update the compatibility table huboard milestone order
0
340,365
10,271,233,953
IssuesEvent
2019-08-23 13:40:26
infor-design/enterprise
https://api.github.com/repos/infor-design/enterprise
closed
Editor: Event callback for mode change - Visual to HTML & vice versa
[3] priority: high type: enhancement :sparkles:
**Description** We need to have event callback when user changes mode from HTML to Visual or vice versa. We would like to do some transformations on HTML code before it is shown in the Visual mode. We would also like to remove those transformations when user switches back to HTML code. To support this, we would like to have call backs where we can listen to mode change events. For example, User writes following in HTML mode ``` <p> Hello {{customeName}}, Thank you for ordering. {{#if selectedForSurvey }} We would like to hear your feedback. Please fill up this survey form. {{/if}} </p> ``` When he switches to Visual mode, we would like to show the user with actual values for variables used in HTML code. To perform these transformations, we would like to have some callbacks to listen to mode change events. We need this in our Angular project. So `enterprise-ng` also needs the fix. **Possible Solution** We need some callback functions like below. The name suggested are for demonstration purpose only. You can choose whatever names are more suitable for ``` this.editor.onBeforePreview( () => { this.editorElementRef.nativeHTML = transform(this.editorElementRef.nativeHTML); } ); ``` ``` this.editor.onBeforeHTML( () => { this.editorElementRef.nativeHTML = reverseTransform(this.editorElementRef.nativeHTML); } ); ``` **Additional context** We would appreciate if this issue can fixed on a high priority basis. We won't be able to show proper preview to our customers without supporting the transformations on HTML code.
1.0
Editor: Event callback for mode change - Visual to HTML & vice versa - **Description** We need to have event callback when user changes mode from HTML to Visual or vice versa. We would like to do some transformations on HTML code before it is shown in the Visual mode. We would also like to remove those transformations when user switches back to HTML code. To support this, we would like to have call backs where we can listen to mode change events. For example, User writes following in HTML mode ``` <p> Hello {{customeName}}, Thank you for ordering. {{#if selectedForSurvey }} We would like to hear your feedback. Please fill up this survey form. {{/if}} </p> ``` When he switches to Visual mode, we would like to show the user with actual values for variables used in HTML code. To perform these transformations, we would like to have some callbacks to listen to mode change events. We need this in our Angular project. So `enterprise-ng` also needs the fix. **Possible Solution** We need some callback functions like below. The name suggested are for demonstration purpose only. You can choose whatever names are more suitable for ``` this.editor.onBeforePreview( () => { this.editorElementRef.nativeHTML = transform(this.editorElementRef.nativeHTML); } ); ``` ``` this.editor.onBeforeHTML( () => { this.editorElementRef.nativeHTML = reverseTransform(this.editorElementRef.nativeHTML); } ); ``` **Additional context** We would appreciate if this issue can fixed on a high priority basis. We won't be able to show proper preview to our customers without supporting the transformations on HTML code.
non_test
editor event callback for mode change visual to html vice versa description we need to have event callback when user changes mode from html to visual or vice versa we would like to do some transformations on html code before it is shown in the visual mode we would also like to remove those transformations when user switches back to html code to support this we would like to have call backs where we can listen to mode change events for example user writes following in html mode hello customename thank you for ordering if selectedforsurvey we would like to hear your feedback please fill up this survey form if when he switches to visual mode we would like to show the user with actual values for variables used in html code to perform these transformations we would like to have some callbacks to listen to mode change events we need this in our angular project so enterprise ng also needs the fix possible solution we need some callback functions like below the name suggested are for demonstration purpose only you can choose whatever names are more suitable for this editor onbeforepreview this editorelementref nativehtml transform this editorelementref nativehtml this editor onbeforehtml this editorelementref nativehtml reversetransform this editorelementref nativehtml additional context we would appreciate if this issue can fixed on a high priority basis we won t be able to show proper preview to our customers without supporting the transformations on html code
0
528,548
15,369,450,739
IssuesEvent
2021-03-02 07:22:57
urbit/landscape
https://api.github.com/repos/urbit/landscape
closed
chat: mention punctuation handling renders !~sampel-palnet as ~sampel-palnet!
priority: 3
**Describe the bug** punctuation preffixed to a mention is rendered suffixed to the mention `!!!!!!!!!~sampel-palnet` becomes `~sampel-palnet!!!!!!!!!` and `!!!~sampel-palnet;;;` becomes `~sampel-palnet!!!;;;` **To Reproduce** Steps to reproduce the behavior: 1. prefix punctuation to mention 2. observe said punctuation now suffixed **Expected behavior** punctuation should stay where it belongs **Desktop (please complete the following information):** - OS: arch linux - Browser qutebrowser - Base hash eljue
1.0
chat: mention punctuation handling renders !~sampel-palnet as ~sampel-palnet! - **Describe the bug** punctuation preffixed to a mention is rendered suffixed to the mention `!!!!!!!!!~sampel-palnet` becomes `~sampel-palnet!!!!!!!!!` and `!!!~sampel-palnet;;;` becomes `~sampel-palnet!!!;;;` **To Reproduce** Steps to reproduce the behavior: 1. prefix punctuation to mention 2. observe said punctuation now suffixed **Expected behavior** punctuation should stay where it belongs **Desktop (please complete the following information):** - OS: arch linux - Browser qutebrowser - Base hash eljue
non_test
chat mention punctuation handling renders sampel palnet as sampel palnet describe the bug punctuation preffixed to a mention is rendered suffixed to the mention sampel palnet becomes sampel palnet and sampel palnet becomes sampel palnet to reproduce steps to reproduce the behavior prefix punctuation to mention observe said punctuation now suffixed expected behavior punctuation should stay where it belongs desktop please complete the following information os arch linux browser qutebrowser base hash eljue
0
740,196
25,740,173,244
IssuesEvent
2022-12-08 05:14:20
googleapis/nodejs-ai-platform
https://api.github.com/repos/googleapis/nodejs-ai-platform
closed
AI platform get training pipeline: should get the specified training pipeline failed
type: bug priority: p1 flakybot: issue api: vertex-ai
Note: #434 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 04f7c858217f1a3ce7b1072c7bf8946d39947532 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2c1b5aae-a46c-4a7a-a474-79da01e686eb), [Sponge](http://sponge2/2c1b5aae-a46c-4a7a-a474-79da01e686eb) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. Error: Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/get-training-pipeline.test.js:23:28) at Context.<anonymous> (test/get-training-pipeline.test.js:31:20) at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
AI platform get training pipeline: should get the specified training pipeline failed - Note: #434 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 04f7c858217f1a3ce7b1072c7bf8946d39947532 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2c1b5aae-a46c-4a7a-a474-79da01e686eb), [Sponge](http://sponge2/2c1b5aae-a46c-4a7a-a474-79da01e686eb) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. Error: Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/get-training-pipeline.test.js:23:28) at Context.<anonymous> (test/get-training-pipeline.test.js:31:20) at processImmediate (internal/timers.js:461:21)</pre></details>
non_test
ai platform get training pipeline should get the specified training pipeline failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output command failed node get training pipeline js undefined undefined permission denied permission denied consumer project undefined has been suspended error command failed node get training pipeline js undefined undefined permission denied permission denied consumer project undefined has been suspended at checkexecsyncerror child process js at object execsync child process js at execsync test get training pipeline test js at context test get training pipeline test js at processimmediate internal timers js
0
17,592
3,012,747,540
IssuesEvent
2015-07-29 02:09:31
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
[CLOSED] Case start fails
auto-migrated Priority-Critical Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Jul 27, 2015 at 03:21 GMT_ _Originally opened as https://github.com/adamsmj/yawl/issues/41_ ---- ``` Starting a case for the attached example yields the following error message: "Unsuccessful case start:Unexpected failure from launchCase (see log for details". Task A has the RRR strategy, the other task is a timer task. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 22 Jul 2008 at 8:06 Attachments: * [new3.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-41/comment-0/new3.xml)
1.0
[CLOSED] Case start fails - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Jul 27, 2015 at 03:21 GMT_ _Originally opened as https://github.com/adamsmj/yawl/issues/41_ ---- ``` Starting a case for the attached example yields the following error message: "Unsuccessful case start:Unexpected failure from launchCase (see log for details". Task A has the RRR strategy, the other task is a timer task. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 22 Jul 2008 at 8:06 Attachments: * [new3.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-41/comment-0/new3.xml)
non_test
case start fails issue by monday jul at gmt originally opened as starting a case for the attached example yields the following error message unsuccessful case start unexpected failure from launchcase see log for details task a has the rrr strategy the other task is a timer task original issue reported on code google com by arthurte gmail com on jul at attachments
0
343,713
30,685,508,790
IssuesEvent
2023-07-26 12:06:03
sarahsporck/Migration-Test-Repository
https://api.github.com/repos/sarahsporck/Migration-Test-Repository
closed
IGAPP-240: Test module wohnen
Native Test
# Max Ammann - 19.3.2019, 18:16:09 None **Environment**: - **Linked issues:**
1.0
IGAPP-240: Test module wohnen - # Max Ammann - 19.3.2019, 18:16:09 None **Environment**: - **Linked issues:**
test
igapp test module wohnen max ammann none environment linked issues
1
204,061
15,398,713,506
IssuesEvent
2021-03-04 00:36:50
nucleus-security/Test-repo
https://api.github.com/repos/nucleus-security/Test-repo
opened
Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [High] - CentOS Security Update for kernel Security Update (CESA-2019:1169)
Testing
Source: QUALYS Finding Description: CentOS has released security update for kernel security update to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6 Impact: N/A</p> Target(s): Asset name: 45.55.254.143 IP: 45.55.254.143 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">centos 6</a> for updates and patch information. <p>Patch:<br /> Following are links for downloading patches to fix the vulnerabilities: </p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">CESA-2019:1169: centos 6</a></p> References: ID:256612 CVE:CVE-2018-12130,CVE-2018-12126,CVE-2018-12127,CVE-2019-11091 Category:CentOS PCI Flagged:1 Vendor References:CESA-2019:1169 centos 6 Severity: High Date Discovered: 2020-01-07 14:35:48 Nucleus Notification Rules Triggered: GitHub Rule Project Name: Ticketing Rules now apply to all vulnerabilities
1.0
Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [High] - CentOS Security Update for kernel Security Update (CESA-2019:1169) - Source: QUALYS Finding Description: CentOS has released security update for kernel security update to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6 Impact: N/A</p> Target(s): Asset name: 45.55.254.143 IP: 45.55.254.143 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">centos 6</a> for updates and patch information. <p>Patch:<br /> Following are links for downloading patches to fix the vulnerabilities: </p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">CESA-2019:1169: centos 6</a></p> References: ID:256612 CVE:CVE-2018-12130,CVE-2018-12126,CVE-2018-12127,CVE-2019-11091 Category:CentOS PCI Flagged:1 Vendor References:CESA-2019:1169 centos 6 Severity: High Date Discovered: 2020-01-07 14:35:48 Nucleus Notification Rules Triggered: GitHub Rule Project Name: Ticketing Rules now apply to all vulnerabilities
test
nucleus project ticketing rules now apply to all vulnerabilities centos security update for kernel security update cesa source qualys finding description centos has released security update for kernel security update to fix the vulnerabilities affected products centos impact n a target s asset name ip solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory for updates and patch information patch following are links for downloading patches to fix the vulnerabilities references id cve cve cve cve cve category centos pci flagged vendor references cesa centos severity high date discovered nucleus notification rules triggered github rule project name ticketing rules now apply to all vulnerabilities
1
113,054
24,356,999,738
IssuesEvent
2022-10-03 08:22:46
julz0815/test-action
https://api.github.com/repos/julz0815/test-action
opened
Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection') [VID:7]
VeracodeFlaw: Very High Veracode Policy Scan
https://github.com/julz0815/test-action/blob/1b75779c50b856e0bd6418d47c0aa10cfa479423/services/admin.service.js#L1-L11 **Filename:** admin.service.js **Line:** 6 **CWE:** 78 (Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection')) <span>This call to util.exec() contains a command injection flaw. The argument to the function is constructed using untrusted input. If an attacker is allowed to specify all or part of the command, it may be possible to execute commands on the server with the privileges of the executing process. The level of exposure depends on the effectiveness of input validation routines, if any. </span> <span>Validate all untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible. When using blocklists, be sure that the sanitizing routine performs a sufficient number of iterations to remove all instances of disallowed characters. Most APIs that execute system commands also have a "safe" version of the method that takes an array of strings as input rather than a single string, which protects against some forms of command injection.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/78.html">CWE</a> <a href="https://owasp.org/www-community/attacks/Command_Injection">OWASP</a></span>
2.0
Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection') [VID:7] - https://github.com/julz0815/test-action/blob/1b75779c50b856e0bd6418d47c0aa10cfa479423/services/admin.service.js#L1-L11 **Filename:** admin.service.js **Line:** 6 **CWE:** 78 (Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection')) <span>This call to util.exec() contains a command injection flaw. The argument to the function is constructed using untrusted input. If an attacker is allowed to specify all or part of the command, it may be possible to execute commands on the server with the privileges of the executing process. The level of exposure depends on the effectiveness of input validation routines, if any. </span> <span>Validate all untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible. When using blocklists, be sure that the sanitizing routine performs a sufficient number of iterations to remove all instances of disallowed characters. Most APIs that execute system commands also have a "safe" version of the method that takes an array of strings as input rather than a single string, which protects against some forms of command injection.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/78.html">CWE</a> <a href="https://owasp.org/www-community/attacks/Command_Injection">OWASP</a></span>
non_test
improper neutralization of special elements used in an os command os command injection command or argument injection filename admin service js line cwe improper neutralization of special elements used in an os command os command injection command or argument injection this call to util exec contains a command injection flaw the argument to the function is constructed using untrusted input if an attacker is allowed to specify all or part of the command it may be possible to execute commands on the server with the privileges of the executing process the level of exposure depends on the effectiveness of input validation routines if any validate all untrusted input to ensure that it conforms to the expected format using centralized data validation routines when possible when using blocklists be sure that the sanitizing routine performs a sufficient number of iterations to remove all instances of disallowed characters most apis that execute system commands also have a safe version of the method that takes an array of strings as input rather than a single string which protects against some forms of command injection references a href a href
0
660,197
21,956,727,160
IssuesEvent
2022-05-24 12:45:49
EddieHubCommunity/LinkFree
https://api.github.com/repos/EddieHubCommunity/LinkFree
closed
[BUG] Copy to clipboard in the Share to social media icons is not working
🕹 aspect: interface 🛠 goal: fix 🟧 priority: high
### Description The following button is not working in the `share to social media` - copy to clip board I tried clicking the button few times and it does not seem to copy the contents of the link. Instead it selects the contents of the description in the profile. I am using BRAVE BROWSER. At the time of this issue the linkfree is at the v0.58.3 ### Screenshots ![ezgif-3-9210861ce8](https://user-images.githubusercontent.com/49341969/148681738-de805239-16fb-4c47-bb8c-c4478dd12845.gif) ### Additional information _No response_
1.0
[BUG] Copy to clipboard in the Share to social media icons is not working - ### Description The following button is not working in the `share to social media` - copy to clip board I tried clicking the button few times and it does not seem to copy the contents of the link. Instead it selects the contents of the description in the profile. I am using BRAVE BROWSER. At the time of this issue the linkfree is at the v0.58.3 ### Screenshots ![ezgif-3-9210861ce8](https://user-images.githubusercontent.com/49341969/148681738-de805239-16fb-4c47-bb8c-c4478dd12845.gif) ### Additional information _No response_
non_test
copy to clipboard in the share to social media icons is not working description the following button is not working in the share to social media copy to clip board i tried clicking the button few times and it does not seem to copy the contents of the link instead it selects the contents of the description in the profile i am using brave browser at the time of this issue the linkfree is at the screenshots additional information no response
0
14,484
3,407,935,839
IssuesEvent
2015-12-04 07:32:27
thedaydreamer/SampleRepo
https://api.github.com/repos/thedaydreamer/SampleRepo
opened
Awesome People
ZipTest
Awesome people are always awesome 1. point one 2. point two You can view a snapshot of the issue using following link : http://webconverger.org/img/lollipop.png
1.0
Awesome People - Awesome people are always awesome 1. point one 2. point two You can view a snapshot of the issue using following link : http://webconverger.org/img/lollipop.png
test
awesome people awesome people are always awesome point one point two you can view a snapshot of the issue using following link
1
338,770
30,320,405,811
IssuesEvent
2023-07-10 18:46:32
ita-social-projects/StreetCode
https://api.github.com/repos/ita-social-projects/StreetCode
opened
Admin [Interesting facts block] Verify if Admin can add no more than 62 symbols to title field
Test case
**Priority:*hight* **Description**: Test case verifies if Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working **Preconditions** 1. Go to http://185.230.138.173/admin-panel 2. Login as 'adminStreetcode ' , password - 'pH2603VkN4d' 3. Go to the Interesting facts block - WOW - факти **Steps to reproduce** 1. Click on "+" button 2. Add the text to the field "Заголовок" 3. Verify if Counter for "symbols left" is displayed and correctly counts all symbols in the text including spaces 4. Add the text with more than 62 symbols to the field "Заголовок" 5. Сheck that admin cannot add text longer than 62 characters in this field, and this title field only allows to add 62 characters. **Expected result** Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working **User story:** #123 **Related bug:**
1.0
Admin [Interesting facts block] Verify if Admin can add no more than 62 symbols to title field - **Priority:*hight* **Description**: Test case verifies if Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working **Preconditions** 1. Go to http://185.230.138.173/admin-panel 2. Login as 'adminStreetcode ' , password - 'pH2603VkN4d' 3. Go to the Interesting facts block - WOW - факти **Steps to reproduce** 1. Click on "+" button 2. Add the text to the field "Заголовок" 3. Verify if Counter for "symbols left" is displayed and correctly counts all symbols in the text including spaces 4. Add the text with more than 62 symbols to the field "Заголовок" 5. Сheck that admin cannot add text longer than 62 characters in this field, and this title field only allows to add 62 characters. **Expected result** Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working **User story:** #123 **Related bug:**
test
admin verify if admin can add no more than symbols to title field priority hight description test case verifies if admin can add no more than symbols to title field and counter for symbols left is displayed and working preconditions go to login as adminstreetcode password go to the interesting facts block wow факти steps to reproduce click on button add the text to the field заголовок verify if counter for symbols left is displayed and correctly counts all symbols in the text including spaces add the text with more than symbols to the field заголовок сheck that admin cannot add text longer than characters in this field and this title field only allows to add characters expected result admin can add no more than symbols to title field and counter for symbols left is displayed and working user story related bug
1
62,690
6,803,385,674
IssuesEvent
2017-11-03 00:36:32
NetsBlox/NetsBlox
https://api.github.com/repos/NetsBlox/NetsBlox
closed
newly created projects (programmatically) cannot be published
bug minor testing
If you create a project programmatically, it cannot be saved as it relies on accessing a non-standard field only available when reading from the database (deprecated and should not be used). This is also a problem exposed by #1637
1.0
newly created projects (programmatically) cannot be published - If you create a project programmatically, it cannot be saved as it relies on accessing a non-standard field only available when reading from the database (deprecated and should not be used). This is also a problem exposed by #1637
test
newly created projects programmatically cannot be published if you create a project programmatically it cannot be saved as it relies on accessing a non standard field only available when reading from the database deprecated and should not be used this is also a problem exposed by
1
117,120
17,410,187,498
IssuesEvent
2021-08-03 11:18:00
tamirdahan/keycloak
https://api.github.com/repos/tamirdahan/keycloak
closed
WS-2017-0116 (Medium) detected in angular-v1.2.32 - autoclosed
security vulnerability
## WS-2017-0116 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-v1.2.32</b></p></summary> <p></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json</p> <p>Path to vulnerable library: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower_components/angular/.bower.json</p> <p> Dependency Hierarchy: - :x: **angular-v1.2.32** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The use element can reference external svg's (same origin) and can include xlink javascript urls or foreign object that can execute xss. <p>Publish Date: 2015-12-05 <p>URL: <a href=https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8>WS-2017-0116</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/pull/13453">https://github.com/angular/angular.js/pull/13453</a></p> <p>Release Date: 2015-12-05</p> <p>Fix Resolution: angular - 1.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Bower","packageName":"angular","packageVersion":"v1.2.32","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json"],"isTransitiveDependency":false,"dependencyTree":"angular:v1.2.32","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0116","vulnerabilityDetails":"The use element can reference external svg\u0027s (same origin) and can include xlink javascript urls or foreign object that can execute xss.","vulnerabilityUrl":"https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8","cvss3Severity":"medium","cvss3Score":"5.8","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2017-0116 (Medium) detected in angular-v1.2.32 - autoclosed - ## WS-2017-0116 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-v1.2.32</b></p></summary> <p></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json</p> <p>Path to vulnerable library: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower_components/angular/.bower.json</p> <p> Dependency Hierarchy: - :x: **angular-v1.2.32** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The use element can reference external svg's (same origin) and can include xlink javascript urls or foreign object that can execute xss. <p>Publish Date: 2015-12-05 <p>URL: <a href=https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8>WS-2017-0116</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/pull/13453">https://github.com/angular/angular.js/pull/13453</a></p> <p>Release Date: 2015-12-05</p> <p>Fix Resolution: angular - 1.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Bower","packageName":"angular","packageVersion":"v1.2.32","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json"],"isTransitiveDependency":false,"dependencyTree":"angular:v1.2.32","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0116","vulnerabilityDetails":"The use element can reference external svg\u0027s (same origin) and can include xlink javascript urls or foreign object that can execute xss.","vulnerabilityUrl":"https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8","cvss3Severity":"medium","cvss3Score":"5.8","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
ws medium detected in angular autoclosed ws medium severity vulnerability vulnerable library angular path to dependency file keycloak themes src main resources theme keycloak common resources node modules autofill event bower json path to vulnerable library keycloak themes src main resources theme keycloak common resources node modules autofill event bower components angular bower json dependency hierarchy x angular vulnerable library found in base branch master vulnerability details the use element can reference external svg s same origin and can include xlink javascript urls or foreign object that can execute xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angular isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree angular isminimumfixversionavailable true minimumfixversion angular basebranches vulnerabilityidentifier ws vulnerabilitydetails the use element can reference external svg same origin and can include xlink javascript urls or foreign object that can execute xss vulnerabilityurl
0
21,597
10,666,994,584
IssuesEvent
2019-10-19 08:49:01
OSWeekends/guilds.osweekends.com
https://api.github.com/repos/OSWeekends/guilds.osweekends.com
opened
WS-2019-0064 (High) detected in handlebars-4.0.11.tgz
security vulnerability
## WS-2019-0064 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.11.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/guilds.osweekends.com/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/guilds.osweekends.com/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - pillars-0.7.1.tgz (Root Library) - templated-0.3.9.tgz - :x: **handlebars-4.0.11.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/OSWeekends/guilds.osweekends.com/commit/7c3567c7b8d78dce0fcb947eca5db3cca195eb19">7c3567c7b8d78dce0fcb947eca5db3cca195eb19</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server. <p>Publish Date: 2019-04-30 <p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0064 (High) detected in handlebars-4.0.11.tgz - ## WS-2019-0064 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.11.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/guilds.osweekends.com/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/guilds.osweekends.com/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - pillars-0.7.1.tgz (Root Library) - templated-0.3.9.tgz - :x: **handlebars-4.0.11.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/OSWeekends/guilds.osweekends.com/commit/7c3567c7b8d78dce0fcb947eca5db3cca195eb19">7c3567c7b8d78dce0fcb947eca5db3cca195eb19</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server. <p>Publish Date: 2019-04-30 <p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws high detected in handlebars tgz ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file tmp ws scm guilds osweekends com package json path to vulnerable library tmp ws scm guilds osweekends com node modules handlebars package json dependency hierarchy pillars tgz root library templated tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details versions of handlebars prior to are vulnerable to prototype pollution templates may alter an objects prototype thus allowing an attacker to execute arbitrary code on the server publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
154,105
13,537,449,117
IssuesEvent
2020-09-16 10:31:08
jolocom/jolo-did-method
https://api.github.com/repos/jolocom/jolo-did-method
closed
Improve documentation
documentation
- Briefly outline the contents of the `packges` folder on the top level of the repo, including references to how and where the separate modules are used. - Document the individual packages.
1.0
Improve documentation - - Briefly outline the contents of the `packges` folder on the top level of the repo, including references to how and where the separate modules are used. - Document the individual packages.
non_test
improve documentation briefly outline the contents of the packges folder on the top level of the repo including references to how and where the separate modules are used document the individual packages
0
160,147
12,504,901,263
IssuesEvent
2020-06-02 09:46:50
Oldes/Rebol-issues
https://api.github.com/repos/Oldes/Rebol-issues
closed
Save/encode png fails under unix platforms
Test.written Type.bug
_Submitted by:_ **johnk** The following works fine under Windows, but fails on Linux x86, Max OS X and Linux ARM (raspberry pi). It seems to be related to encode. Saving to a bmp works correctly as expected. ``` rebol save %test.png make image! [1024x768] encode 'png make image! [1024x768] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2040)** [ Version: r3 master Type: Bug Platform: Linux x86 libc6 Category: Unspecified Reproduce: Always Fixed-in:none ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2040</sup> Comments: --- > **Rebolbot** commented on Sep 30, 2013: _Submitted by:_ **abolka** Also crashes for me on Win32 (using http://www.rebolsource.net/downloads/win32-x86/r3-g4d9840f.exe). --- > **Rebolbot** commented on Sep 30, 2013: _Submitted by:_ **fork** I believe this to be resolved by https://github.com/rebol/rebol/pull/155 --- > **Rebolbot** commented on Sep 30, 2013: _Submitted by:_ **abolka** In the core-tests suite. --- > **Rebolbot** commented on Mar 2, 2014: _Submitted by:_ **abolka** Also addressed by https://github.com/rebol/rebol/pull/194. --- > **Rebolbot** added **Type.bug** and **Test.written** on Jan 12, 2016 --- > **Oldes** commented on May 29, 2019: This one should be closed as it is not an issue anymore. ---
1.0
Save/encode png fails under unix platforms - _Submitted by:_ **johnk** The following works fine under Windows, but fails on Linux x86, Max OS X and Linux ARM (raspberry pi). It seems to be related to encode. Saving to a bmp works correctly as expected. ``` rebol save %test.png make image! [1024x768] encode 'png make image! [1024x768] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2040)** [ Version: r3 master Type: Bug Platform: Linux x86 libc6 Category: Unspecified Reproduce: Always Fixed-in:none ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2040</sup> Comments: --- > **Rebolbot** commented on Sep 30, 2013: _Submitted by:_ **abolka** Also crashes for me on Win32 (using http://www.rebolsource.net/downloads/win32-x86/r3-g4d9840f.exe). --- > **Rebolbot** commented on Sep 30, 2013: _Submitted by:_ **fork** I believe this to be resolved by https://github.com/rebol/rebol/pull/155 --- > **Rebolbot** commented on Sep 30, 2013: _Submitted by:_ **abolka** In the core-tests suite. --- > **Rebolbot** commented on Mar 2, 2014: _Submitted by:_ **abolka** Also addressed by https://github.com/rebol/rebol/pull/194. --- > **Rebolbot** added **Type.bug** and **Test.written** on Jan 12, 2016 --- > **Oldes** commented on May 29, 2019: This one should be closed as it is not an issue anymore. ---
test
save encode png fails under unix platforms submitted by johnk the following works fine under windows but fails on linux max os x and linux arm raspberry pi it seems to be related to encode saving to a bmp works correctly as expected rebol save test png make image encode png make image imported from imported from comments rebolbot commented on sep submitted by abolka also crashes for me on using rebolbot commented on sep submitted by fork i believe this to be resolved by rebolbot commented on sep submitted by abolka in the core tests suite rebolbot commented on mar submitted by abolka also addressed by rebolbot added type bug and test written on jan oldes commented on may this one should be closed as it is not an issue anymore
1
91,693
10,726,867,634
IssuesEvent
2019-10-28 10:20:45
suvajit-sarkar/github-upload
https://api.github.com/repos/suvajit-sarkar/github-upload
opened
Indy: Ansible role to get EIP metadata from AWS
documentation
Prepare an Ansible role which retrieves the EIP allocation IDs for all steward public IPs found in the network.yaml and stores them in an Ansible fact / variable to be used later. The role should also check that the same public IP is specified for all stewards within one organization.
1.0
Indy: Ansible role to get EIP metadata from AWS - Prepare an Ansible role which retrieves the EIP allocation IDs for all steward public IPs found in the network.yaml and stores them in an Ansible fact / variable to be used later. The role should also check that the same public IP is specified for all stewards within one organization.
non_test
indy ansible role to get eip metadata from aws prepare an ansible role which retrieves the eip allocation ids for all steward public ips found in the network yaml and stores them in an ansible fact variable to be used later the role should also check that the same public ip is specified for all stewards within one organization
0
528,145
15,360,844,578
IssuesEvent
2021-03-01 17:24:16
edgedb/edgedb
https://api.github.com/repos/edgedb/edgedb
closed
Migrations aren't propagated to existing connections
bug high priority
- EdgeDB Version:1.0-alpha.8+dev.5386.g6933b454f Steps to Reproduce: 1. Start `edgedb` (repl) and do a query: ``` edgedb> SELECT Counter; error: object type or alias 'default::Counter' does not exist ┌─ query:1:8 │ 1 │ SELECT Counter; │ ^^^^^^^ error ``` 2. Then in another shell do `edgedb migrate` to create the type 3. Then return back to this shell and confirm that error still there (while it shouldn't be): ``` edgedb> SELECT Counter; error: object type or alias 'default::Counter' does not exist ┌─ query:1:8 │ 1 │ SELECT Counter; │ ^^^^^^^ error ``` 4. Also try and confirm that it works in a new connection: ``` edgedb> SELECT Counter; {} ``` **Note:** this doesn't reproduce with `CREATE TYPE` only with migrations.
1.0
Migrations aren't propagated to existing connections - - EdgeDB Version:1.0-alpha.8+dev.5386.g6933b454f Steps to Reproduce: 1. Start `edgedb` (repl) and do a query: ``` edgedb> SELECT Counter; error: object type or alias 'default::Counter' does not exist ┌─ query:1:8 │ 1 │ SELECT Counter; │ ^^^^^^^ error ``` 2. Then in another shell do `edgedb migrate` to create the type 3. Then return back to this shell and confirm that error still there (while it shouldn't be): ``` edgedb> SELECT Counter; error: object type or alias 'default::Counter' does not exist ┌─ query:1:8 │ 1 │ SELECT Counter; │ ^^^^^^^ error ``` 4. Also try and confirm that it works in a new connection: ``` edgedb> SELECT Counter; {} ``` **Note:** this doesn't reproduce with `CREATE TYPE` only with migrations.
non_test
migrations aren t propagated to existing connections edgedb version alpha dev steps to reproduce start edgedb repl and do a query edgedb select counter error object type or alias default counter does not exist ┌─ query │ │ select counter │ error then in another shell do edgedb migrate to create the type then return back to this shell and confirm that error still there while it shouldn t be edgedb select counter error object type or alias default counter does not exist ┌─ query │ │ select counter │ error also try and confirm that it works in a new connection edgedb select counter note this doesn t reproduce with create type only with migrations
0
184,309
31,854,850,425
IssuesEvent
2023-09-15 06:33:54
opengovsg/FormSG
https://api.github.com/repos/opengovsg/FormSG
closed
Announcement modal
design P4
Why - Inform users of major new features - better than email as users can try it out straight away instead of the notification coming in between possibly other work they have to attend to Qns - Is the v3 "What's New" / notifications design sufficient? Implementation - Can check out Postman's https://github.com/opengovsg/postmangovsg/pull/920
1.0
Announcement modal - Why - Inform users of major new features - better than email as users can try it out straight away instead of the notification coming in between possibly other work they have to attend to Qns - Is the v3 "What's New" / notifications design sufficient? Implementation - Can check out Postman's https://github.com/opengovsg/postmangovsg/pull/920
non_test
announcement modal why inform users of major new features better than email as users can try it out straight away instead of the notification coming in between possibly other work they have to attend to qns is the what s new notifications design sufficient implementation can check out postman s
0
12,341
3,266,862,989
IssuesEvent
2015-10-22 22:46:45
NorthBridge/nexus-community
https://api.github.com/repos/NorthBridge/nexus-community
closed
Nexus Web Meet timeout oid query param value incorrect
ready to test
Should show a message that indicates user session has timed out. Right now message says they logged out.
1.0
Nexus Web Meet timeout oid query param value incorrect - Should show a message that indicates user session has timed out. Right now message says they logged out.
test
nexus web meet timeout oid query param value incorrect should show a message that indicates user session has timed out right now message says they logged out
1
289,502
24,993,866,330
IssuesEvent
2022-11-02 21:27:25
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
acctest: Failing acceptance tests
tests
Acceptance tests are failing in the `acctest` package: ``` === RUN TestAccProvider_DefaultTagsTags_none === PAUSE TestAccProvider_DefaultTagsTags_none === CONT TestAccProvider_DefaultTagsTags_none 2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Starting TestCase: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1 2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Calling TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1 2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Called TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1 2022-11-02T05:10:14.038Z [WARN] sdk.helper_resource: Error running Terraform CLI command: test_terraform_path=/opt/teamcity-agent/temp/buildTmp/plugintest-terraform3209408513/terraform test_working_directory=/opt/teamcity-agent/temp/buildTmp/plugintest3846192360 error= | exit status 1 | | Error: Invalid data source | | on terraform_plugin_test.tf line 2, in data "aws_region" "provider_test": | 2: data "aws_region" "provider_test" {} | | The provider hashicorp/aws does not support data source "aws_region". | | Error: Invalid data source | | on terraform_plugin_test.tf line 5, in data "aws_service" "provider_test": | 5: data "aws_service" "provider_test" { | | The provider hashicorp/aws does not support data source "aws_service". ``` Relates https://github.com/hashicorp/terraform-provider-aws/pull/27221.
1.0
acctest: Failing acceptance tests - Acceptance tests are failing in the `acctest` package: ``` === RUN TestAccProvider_DefaultTagsTags_none === PAUSE TestAccProvider_DefaultTagsTags_none === CONT TestAccProvider_DefaultTagsTags_none 2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Starting TestCase: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1 2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Calling TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1 2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Called TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1 2022-11-02T05:10:14.038Z [WARN] sdk.helper_resource: Error running Terraform CLI command: test_terraform_path=/opt/teamcity-agent/temp/buildTmp/plugintest-terraform3209408513/terraform test_working_directory=/opt/teamcity-agent/temp/buildTmp/plugintest3846192360 error= | exit status 1 | | Error: Invalid data source | | on terraform_plugin_test.tf line 2, in data "aws_region" "provider_test": | 2: data "aws_region" "provider_test" {} | | The provider hashicorp/aws does not support data source "aws_region". | | Error: Invalid data source | | on terraform_plugin_test.tf line 5, in data "aws_service" "provider_test": | 5: data "aws_service" "provider_test" { | | The provider hashicorp/aws does not support data source "aws_service". ``` Relates https://github.com/hashicorp/terraform-provider-aws/pull/27221.
test
acctest failing acceptance tests acceptance tests are failing in the acctest package run testaccprovider defaulttagstags none pause testaccprovider defaulttagstags none cont testaccprovider defaulttagstags none sdk helper resource starting testcase test name testaccprovider defaulttagstags none test step number sdk helper resource calling testcase precheck test name testaccprovider defaulttagstags none test step number sdk helper resource called testcase precheck test name testaccprovider defaulttagstags none test step number sdk helper resource error running terraform cli command test terraform path opt teamcity agent temp buildtmp plugintest terraform test working directory opt teamcity agent temp buildtmp error exit status error invalid data source on terraform plugin test tf line in data aws region provider test data aws region provider test the provider hashicorp aws does not support data source aws region error invalid data source on terraform plugin test tf line in data aws service provider test data aws service provider test the provider hashicorp aws does not support data source aws service relates
1
22,834
20,269,206,753
IssuesEvent
2022-02-15 14:50:01
firemodels/fds
https://api.github.com/repos/firemodels/fds
closed
Output A and E in TGA_ANALYSIS mode
Usability
Simulations with TGA_ANALYSIS = .TRUE. usually start with REFERENCE_TEMPERATURE etc (first order), but if user wants to see the resulting A and E, and continue by specifying / modifying them directly, it would be convenient to see the calculated values from somewhere. Normally, FDS writes A and E in .out file, under MATERIAL definitions. But TGA_ANALYSIS -mode does not produce .out file. Finding A and E requires removing TGA_ANALYSIS = .TRUE. -line, and making sure the file runs otherwise. Unnecessary steps in the process. How about writing A and E directly to LU_ERR before the 'STOP: FDS performed.... ' -line?
True
Output A and E in TGA_ANALYSIS mode - Simulations with TGA_ANALYSIS = .TRUE. usually start with REFERENCE_TEMPERATURE etc (first order), but if user wants to see the resulting A and E, and continue by specifying / modifying them directly, it would be convenient to see the calculated values from somewhere. Normally, FDS writes A and E in .out file, under MATERIAL definitions. But TGA_ANALYSIS -mode does not produce .out file. Finding A and E requires removing TGA_ANALYSIS = .TRUE. -line, and making sure the file runs otherwise. Unnecessary steps in the process. How about writing A and E directly to LU_ERR before the 'STOP: FDS performed.... ' -line?
non_test
output a and e in tga analysis mode simulations with tga analysis true usually start with reference temperature etc first order but if user wants to see the resulting a and e and continue by specifying modifying them directly it would be convenient to see the calculated values from somewhere normally fds writes a and e in out file under material definitions but tga analysis mode does not produce out file finding a and e requires removing tga analysis true line and making sure the file runs otherwise unnecessary steps in the process how about writing a and e directly to lu err before the stop fds performed line
0
192,259
14,612,792,712
IssuesEvent
2020-12-22 06:56:24
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
YanshuoH/youkonger-backend: dao/event_date_test.go; 4 LoC
fresh test tiny
Found a possible issue in [YanshuoH/youkonger-backend](https://www.github.com/YanshuoH/youkonger-backend) at [dao/event_date_test.go](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call at line 53 may store a reference to ep [Click here to see the code in its original context.](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54) <details> <summary>Click here to show the 4 line(s) of Go which triggered the analyzer.</summary> ```go for _, ep := range epToInserts { ep.EventDateID = edToInsert.ID Expect(Conn.Create(&ep).Error).ToNot(HaveOccurred()) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: dcd1175b9976855007d9c3cdc9c70ec685dd5f14
1.0
YanshuoH/youkonger-backend: dao/event_date_test.go; 4 LoC - Found a possible issue in [YanshuoH/youkonger-backend](https://www.github.com/YanshuoH/youkonger-backend) at [dao/event_date_test.go](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call at line 53 may store a reference to ep [Click here to see the code in its original context.](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54) <details> <summary>Click here to show the 4 line(s) of Go which triggered the analyzer.</summary> ```go for _, ep := range epToInserts { ep.EventDateID = edToInsert.ID Expect(Conn.Create(&ep).Error).ToNot(HaveOccurred()) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: dcd1175b9976855007d9c3cdc9c70ec685dd5f14
test
yanshuoh youkonger backend dao event date test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call at line may store a reference to ep click here to show the line s of go which triggered the analyzer go for ep range eptoinserts ep eventdateid edtoinsert id expect conn create ep error tonot haveoccurred leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
179,189
13,850,131,129
IssuesEvent
2020-10-15 00:20:11
xbmc/metadata.tvshows.themoviedb.org.python
https://api.github.com/repos/xbmc/metadata.tvshows.themoviedb.org.python
closed
Episode Titles not scraped with fallback language
bug testing
Issue reported here for the XML scraper also affects this Python scraper... https://github.com/xbmc/repo-scrapers/issues/225
1.0
Episode Titles not scraped with fallback language - Issue reported here for the XML scraper also affects this Python scraper... https://github.com/xbmc/repo-scrapers/issues/225
test
episode titles not scraped with fallback language issue reported here for the xml scraper also affects this python scraper
1
182,379
14,117,051,253
IssuesEvent
2020-11-08 07:07:42
fabioturazzidouglas/Hackermen_Project
https://api.github.com/repos/fabioturazzidouglas/Hackermen_Project
closed
JUnit Booking System Testing
Testing
Tests the iteration between all models (Users, Bookings, Rooms, RoomTypes) to create bookings
1.0
JUnit Booking System Testing - Tests the iteration between all models (Users, Bookings, Rooms, RoomTypes) to create bookings
test
junit booking system testing tests the iteration between all models users bookings rooms roomtypes to create bookings
1
126,230
10,413,013,799
IssuesEvent
2019-09-13 17:26:35
futest-test/fu
https://api.github.com/repos/futest-test/fu
opened
Vulnerability - Cross site scripting
FoundByAcunetix360 FuTest
**URL:** http://php.testsparker.com/ **Name:** Cross site scripting **Severity:** High You can see vulnerability details from the link below. http://ec2-18-194-173-226.eu-central-1.compute.amazonaws.com/vulnerabilities/detail/df6c902af25242c98a98aac602aa83ff
1.0
Vulnerability - Cross site scripting - **URL:** http://php.testsparker.com/ **Name:** Cross site scripting **Severity:** High You can see vulnerability details from the link below. http://ec2-18-194-173-226.eu-central-1.compute.amazonaws.com/vulnerabilities/detail/df6c902af25242c98a98aac602aa83ff
test
vulnerability cross site scripting url name cross site scripting severity high you can see vulnerability details from the link below
1
548,545
16,066,511,475
IssuesEvent
2021-04-23 20:03:16
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Cannot Delete Last workflow survey question
component:ui priority:high state:in_progress type:bug
##### ISSUE TYPE - Bug Report ##### SUMMARY <img width="1483" alt="Screen Shot 2020-07-20 at 1 25 35 PM" src="https://user-images.githubusercontent.com/9889020/87967264-93ac5b80-ca8c-11ea-8050-66400bb6a019.png"> If you create a survey question for a workflow job template and then try to delete that question you get the above error. It appears as if we're trying to delete the question from `/api/v2/job_templates/:id/survey_spec` instead of `/api/v2/workflow_job_templates/:id/survey_spec` ##### STEPS TO REPRODUCE 1. Create a survey for a WFJT 2. Enable surveys for WFJT 3. Select Survey 4. Attempt to delete Survey Note: 404 also 1. Create many surveys for a WFJT 2. Enable surveys for WFJT 3. Select All surveys 4. Attempt to delete all surveys Note: 404
1.0
Cannot Delete Last workflow survey question - ##### ISSUE TYPE - Bug Report ##### SUMMARY <img width="1483" alt="Screen Shot 2020-07-20 at 1 25 35 PM" src="https://user-images.githubusercontent.com/9889020/87967264-93ac5b80-ca8c-11ea-8050-66400bb6a019.png"> If you create a survey question for a workflow job template and then try to delete that question you get the above error. It appears as if we're trying to delete the question from `/api/v2/job_templates/:id/survey_spec` instead of `/api/v2/workflow_job_templates/:id/survey_spec` ##### STEPS TO REPRODUCE 1. Create a survey for a WFJT 2. Enable surveys for WFJT 3. Select Survey 4. Attempt to delete Survey Note: 404 also 1. Create many surveys for a WFJT 2. Enable surveys for WFJT 3. Select All surveys 4. Attempt to delete all surveys Note: 404
non_test
cannot delete last workflow survey question issue type bug report summary img width alt screen shot at pm src if you create a survey question for a workflow job template and then try to delete that question you get the above error it appears as if we re trying to delete the question from api job templates id survey spec instead of api workflow job templates id survey spec steps to reproduce create a survey for a wfjt enable surveys for wfjt select survey attempt to delete survey note also create many surveys for a wfjt enable surveys for wfjt select all surveys attempt to delete all surveys note
0
151,315
12,031,686,994
IssuesEvent
2020-04-13 10:15:58
microsoft/azure-tools-for-java
https://api.github.com/repos/microsoft/azure-tools-for-java
closed
Can't trigger function/admin http function when click 'Trigger Function' button
External Dependency Internal Test function
#### Environment: - OS : Win&Mac - Plugins Version : Azure java toolkit for IntelliJ: 3.34.0 - IDE Version: IU-jre8 & IC-jre11 #### Repro steps: 1. Create a function project, add function/admin auth type http trigger and run deploy. 2. Go to azure explorer, find the function app and right click 'Trigger Function' button. #### Result: ![image](https://user-images.githubusercontent.com/46642771/77063398-959bda80-6a18-11ea-932c-acf72bebd107.png)
1.0
Can't trigger function/admin http function when click 'Trigger Function' button - #### Environment: - OS : Win&Mac - Plugins Version : Azure java toolkit for IntelliJ: 3.34.0 - IDE Version: IU-jre8 & IC-jre11 #### Repro steps: 1. Create a function project, add function/admin auth type http trigger and run deploy. 2. Go to azure explorer, find the function app and right click 'Trigger Function' button. #### Result: ![image](https://user-images.githubusercontent.com/46642771/77063398-959bda80-6a18-11ea-932c-acf72bebd107.png)
test
can t trigger function admin http function when click trigger function button environment: os win mac plugins version azure java toolkit for intellij ide version iu ic repro steps create a function project add function admin auth type http trigger and run deploy go to azure explorer find the function app and right click trigger function button result
1
110,357
13,906,716,633
IssuesEvent
2020-10-20 11:39:36
httpwg/httpbis-issues
https://api.github.com/repos/httpwg/httpbis-issues
opened
Ordering in Upgrade (Trac #445)
Incomplete Migration Migrated from Trac design draft-ietf-httpbis-p1-messaging@tools.ietf.org p1-messaging
Migrated from https://trac.ietf.org/ticket/445 ```json { "status": "closed", "changetime": "2013-09-26T05:56:39", "_ts": "1380174999465154", "description": "p1 section 6.7 defines the Upgrade header, but no where does it say anything about relative preference.\n\nShould we define (or at least allow) for the ordering to be semantically significant? It seems to me that if we end up using this, and there are a few different variants of HTTP/2 (e.g., \"normal\" vs \"mobile\"), it'd be nice to rely on ordering here.", "reporter": "mnot@pobox.com", "cc": "", "resolution": "fixed", "time": "2013-04-20T07:51:45", "component": "p1-messaging", "summary": "Ordering in Upgrade", "priority": "normal", "keywords": "", "milestone": "23", "owner": "draft-ietf-httpbis-p1-messaging@tools.ietf.org", "type": "design", "severity": "In WG Last Call" } ```
1.0
Ordering in Upgrade (Trac #445) - Migrated from https://trac.ietf.org/ticket/445 ```json { "status": "closed", "changetime": "2013-09-26T05:56:39", "_ts": "1380174999465154", "description": "p1 section 6.7 defines the Upgrade header, but no where does it say anything about relative preference.\n\nShould we define (or at least allow) for the ordering to be semantically significant? It seems to me that if we end up using this, and there are a few different variants of HTTP/2 (e.g., \"normal\" vs \"mobile\"), it'd be nice to rely on ordering here.", "reporter": "mnot@pobox.com", "cc": "", "resolution": "fixed", "time": "2013-04-20T07:51:45", "component": "p1-messaging", "summary": "Ordering in Upgrade", "priority": "normal", "keywords": "", "milestone": "23", "owner": "draft-ietf-httpbis-p1-messaging@tools.ietf.org", "type": "design", "severity": "In WG Last Call" } ```
non_test
ordering in upgrade trac migrated from json status closed changetime ts description section defines the upgrade header but no where does it say anything about relative preference n nshould we define or at least allow for the ordering to be semantically significant it seems to me that if we end up using this and there are a few different variants of http e g normal vs mobile it d be nice to rely on ordering here reporter mnot pobox com cc resolution fixed time component messaging summary ordering in upgrade priority normal keywords milestone owner draft ietf httpbis messaging tools ietf org type design severity in wg last call
0
38,548
6,676,677,711
IssuesEvent
2017-10-05 07:14:48
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Python3 doctest compatibility: algorithms IntegrateByComponent to Integration
Component: Documentation Misc: Maintenance
This issue is part of the move to Python3 in the doctests, Algorithms covered in this issue are: - [x] IntegrateByComponent - [x] IntegrateEPP - [x] IntegrateEllipsoids - [x] IntegrateEllipsoidsTwoStep - [x] IntegrateFlux - [x] IntegrateMDHistoWorkspace - [x] IntegratePeakTimeSlices - [x] IntegratePeaksCWSD - [x] IntegratePeaksHybrid - [x] IntegratePeaksMD - [x] IntegratePeaksMDHKL - [x] IntegratePeaksUsingClusters - [x] Integration
1.0
Python3 doctest compatibility: algorithms IntegrateByComponent to Integration - This issue is part of the move to Python3 in the doctests, Algorithms covered in this issue are: - [x] IntegrateByComponent - [x] IntegrateEPP - [x] IntegrateEllipsoids - [x] IntegrateEllipsoidsTwoStep - [x] IntegrateFlux - [x] IntegrateMDHistoWorkspace - [x] IntegratePeakTimeSlices - [x] IntegratePeaksCWSD - [x] IntegratePeaksHybrid - [x] IntegratePeaksMD - [x] IntegratePeaksMDHKL - [x] IntegratePeaksUsingClusters - [x] Integration
non_test
doctest compatibility algorithms integratebycomponent to integration this issue is part of the move to in the doctests algorithms covered in this issue are integratebycomponent integrateepp integrateellipsoids integrateellipsoidstwostep integrateflux integratemdhistoworkspace integratepeaktimeslices integratepeakscwsd integratepeakshybrid integratepeaksmd integratepeaksmdhkl integratepeaksusingclusters integration
0
318,684
27,321,017,404
IssuesEvent
2023-02-24 19:50:52
peviitor-ro/ui-js
https://api.github.com/repos/peviitor-ro/ui-js
closed
[SERP] Different font family for the text "Alătură-te"
bug TestQuality
## Precondition URL: https://beta.peviitor.ro/ Device: Samsung Galaxy S21 Ultra Browser: Chrome Platform: Android 12 ## Steps to Reproduce: ### Step 1 <span style="color:#58b880"> **[Pass]** </span> Open URL in browser #### Expected Result Website is loaded without any error ### Step 2 <span style="color:#58b880"> **[Pass]** </span> Click on “Caută” #### Expected Result The user is redirected to SERP ### Step 3 <span style="color:#ff5538"> **[Fail]** </span> Inspect the font of the text &quot;Alătură-te&quot; #### Expected Result Font is &#x27;SF Pro&#x27; #### Actual Result The actual font is SegoeUI
1.0
[SERP] Different font family for the text "Alătură-te" - ## Precondition URL: https://beta.peviitor.ro/ Device: Samsung Galaxy S21 Ultra Browser: Chrome Platform: Android 12 ## Steps to Reproduce: ### Step 1 <span style="color:#58b880"> **[Pass]** </span> Open URL in browser #### Expected Result Website is loaded without any error ### Step 2 <span style="color:#58b880"> **[Pass]** </span> Click on “Caută” #### Expected Result The user is redirected to SERP ### Step 3 <span style="color:#ff5538"> **[Fail]** </span> Inspect the font of the text &quot;Alătură-te&quot; #### Expected Result Font is &#x27;SF Pro&#x27; #### Actual Result The actual font is SegoeUI
test
different font family for the text alătură te precondition url device samsung galaxy ultra browser chrome platform android steps to reproduce step open url in browser expected result website is loaded without any error step click on “caută” expected result the user is redirected to serp step inspect the font of the text quot alătură te quot expected result font is sf pro actual result the actual font is segoeui
1
118,029
15,216,440,567
IssuesEvent
2021-02-17 15:30:39
wordpress-mobile/WordPress-Android
https://api.github.com/repos/wordpress-mobile/WordPress-Android
opened
Update Activity Log Details UI and Add Backup Download Button/Flow
Jetpack Jetpack Mobile Needs Design [Type] Enhancement
As per the requirements in this post (`pcdRpT-cM`, see `Activity Log: Event Details`), this issue is about: - UI updates on this screen to match design, and - Adding the `Backup Download` button and connecting the button with the new flow. Before | After -------|------ <img width="250" height="500" alt="before" src="https://user-images.githubusercontent.com/9729923/108225084-4092a800-7144-11eb-9458-e2ece9899626.png"> | <img width="250" height="500" alt="after" src="https://user-images.githubusercontent.com/9729923/108225150-53a57800-7144-11eb-88ef-dce217851bd4.png"> On Android, the screen looks a bit different today, comparing the before (or current) of Android with the after (or current) of iOS: Before or Current (Android) | After or Current (iOS) -------|------ <img width="250" height="500" alt="before android" src="https://user-images.githubusercontent.com/9729923/108225656-cf9fc000-7144-11eb-9232-0fc8cb8126a1.png"> | <img width="250" height="500" alt="after ios" src="https://user-images.githubusercontent.com/9729923/108226033-302efd00-7145-11eb-8eaf-883176d46547.png"> @osullivanchris can you please suggest what UI changes need to happen to the UI so that Android get the appropriate after look, and still be comparable with iOS's after look?
1.0
Update Activity Log Details UI and Add Backup Download Button/Flow - As per the requirements in this post (`pcdRpT-cM`, see `Activity Log: Event Details`), this issue is about: - UI updates on this screen to match design, and - Adding the `Backup Download` button and connecting the button with the new flow. Before | After -------|------ <img width="250" height="500" alt="before" src="https://user-images.githubusercontent.com/9729923/108225084-4092a800-7144-11eb-9458-e2ece9899626.png"> | <img width="250" height="500" alt="after" src="https://user-images.githubusercontent.com/9729923/108225150-53a57800-7144-11eb-88ef-dce217851bd4.png"> On Android, the screen looks a bit different today, comparing the before (or current) of Android with the after (or current) of iOS: Before or Current (Android) | After or Current (iOS) -------|------ <img width="250" height="500" alt="before android" src="https://user-images.githubusercontent.com/9729923/108225656-cf9fc000-7144-11eb-9232-0fc8cb8126a1.png"> | <img width="250" height="500" alt="after ios" src="https://user-images.githubusercontent.com/9729923/108226033-302efd00-7145-11eb-8eaf-883176d46547.png"> @osullivanchris can you please suggest what UI changes need to happen to the UI so that Android get the appropriate after look, and still be comparable with iOS's after look?
non_test
update activity log details ui and add backup download button flow as per the requirements in this post pcdrpt cm see activity log event details this issue is about ui updates on this screen to match design and adding the backup download button and connecting the button with the new flow before after img width height alt before src img width height alt after src on android the screen looks a bit different today comparing the before or current of android with the after or current of ios before or current android after or current ios img width height alt before android src img width height alt after ios src osullivanchris can you please suggest what ui changes need to happen to the ui so that android get the appropriate after look and still be comparable with ios s after look
0
4,416
3,369,801,513
IssuesEvent
2015-11-23 12:10:36
mozilla/feedthefox
https://api.github.com/repos/mozilla/feedthefox
closed
Show "code name" on Devices pages
Builds/Devices Discussion UI/UX
It might be good idea to tell people the "code name" of their device at https://firefoxos.mozilla.community/devices/ because they're most likely going to need that bit of information once they click the big green buttons. The ones I know of so far FWIW: Xperia Z3 (leo) Xperia Z3 Compact (aries) Xperia Z2 (sirius) Xperia T2 Ultra (tianchi) Xperia E3 (flamingo)
1.0
Show "code name" on Devices pages - It might be good idea to tell people the "code name" of their device at https://firefoxos.mozilla.community/devices/ because they're most likely going to need that bit of information once they click the big green buttons. The ones I know of so far FWIW: Xperia Z3 (leo) Xperia Z3 Compact (aries) Xperia Z2 (sirius) Xperia T2 Ultra (tianchi) Xperia E3 (flamingo)
non_test
show code name on devices pages it might be good idea to tell people the code name of their device at because they re most likely going to need that bit of information once they click the big green buttons the ones i know of so far fwiw xperia leo xperia compact aries xperia sirius xperia ultra tianchi xperia flamingo
0
28,115
4,365,170,922
IssuesEvent
2016-08-03 09:48:38
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
circleci: failed tests: TestDockerPython
Robot test-failure
The following test appears to have failed: [#20875](https://circleci.com/gh/cockroachdb/cockroach/20875): ``` ^ SyntaxError: invalid syntax I160802 15:54:31.215544 acceptance/cluster/localcluster.go:673 stopping I160802 15:54:31.215794 acceptance/cluster/localcluster.go:573 event stream done, resetting...: net/http: request canceled I160802 15:54:31.215807 acceptance/cluster/localcluster.go:596 events monitor exits --- FAIL: TestDockerPython (20.37s) panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist [recovered] panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist goroutine 2703 [running]: panic(0x14919a0, 0xc8204a3210) /usr/local/go/src/runtime/panic.go:481 +0x3e6 testing.tRunner.func1(0xc8202b0360) /usr/local/go/src/testing/testing.go:467 +0x192 panic(0x14919a0, 0xc8204a3210) /usr/local/go/src/runtime/panic.go:443 +0x4e9 github.com/cockroachdb/cockroach/acceptance/cluster.maybePanic(0x7fca667c1028, 0xc8204a3210) /go/src/github.com/cockroachdb/cockroach/acceptance/cluster/docker.go:166 +0x4b github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).stop(0xc8204641c0) /go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:699 +0x620 github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).AssertAndStop(0xc8204641c0, 0xc8202b0360) /go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:664 +0x53 github.com/cockroachdb/cockroach/acceptance.testDocker(0xc8202b0360, 0x1, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x7fca66789408, 0xc82042eab0) /go/src/github.com/cockroachdb/cockroach/acceptance/util_test.go:406 +0x5bc github.com/cockroachdb/cockroach/acceptance.testDockerSingleNode(0xc8202b0360, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x0, 0x0) ``` Please assign, take a look and update the issue accordingly.
1.0
circleci: failed tests: TestDockerPython - The following test appears to have failed: [#20875](https://circleci.com/gh/cockroachdb/cockroach/20875): ``` ^ SyntaxError: invalid syntax I160802 15:54:31.215544 acceptance/cluster/localcluster.go:673 stopping I160802 15:54:31.215794 acceptance/cluster/localcluster.go:573 event stream done, resetting...: net/http: request canceled I160802 15:54:31.215807 acceptance/cluster/localcluster.go:596 events monitor exits --- FAIL: TestDockerPython (20.37s) panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist [recovered] panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist goroutine 2703 [running]: panic(0x14919a0, 0xc8204a3210) /usr/local/go/src/runtime/panic.go:481 +0x3e6 testing.tRunner.func1(0xc8202b0360) /usr/local/go/src/testing/testing.go:467 +0x192 panic(0x14919a0, 0xc8204a3210) /usr/local/go/src/runtime/panic.go:443 +0x4e9 github.com/cockroachdb/cockroach/acceptance/cluster.maybePanic(0x7fca667c1028, 0xc8204a3210) /go/src/github.com/cockroachdb/cockroach/acceptance/cluster/docker.go:166 +0x4b github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).stop(0xc8204641c0) /go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:699 +0x620 github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).AssertAndStop(0xc8204641c0, 0xc8202b0360) /go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:664 +0x53 github.com/cockroachdb/cockroach/acceptance.testDocker(0xc8202b0360, 0x1, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x7fca66789408, 0xc82042eab0) /go/src/github.com/cockroachdb/cockroach/acceptance/util_test.go:406 +0x5bc github.com/cockroachdb/cockroach/acceptance.testDockerSingleNode(0xc8202b0360, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x0, 0x0) ``` Please assign, take a look and update the issue accordingly.
test
circleci failed tests testdockerpython the following test appears to have failed syntaxerror invalid syntax acceptance cluster localcluster go stopping acceptance cluster localcluster go event stream done resetting net http request canceled acceptance cluster localcluster go events monitor exits fail testdockerpython panic error response from daemon cannot kill container active container for does not exist panic error response from daemon cannot kill container active container for does not exist goroutine panic usr local go src runtime panic go testing trunner usr local go src testing testing go panic usr local go src runtime panic go github com cockroachdb cockroach acceptance cluster maybepanic go src github com cockroachdb cockroach acceptance cluster docker go github com cockroachdb cockroach acceptance cluster localcluster stop go src github com cockroachdb cockroach acceptance cluster localcluster go github com cockroachdb cockroach acceptance cluster localcluster assertandstop go src github com cockroachdb cockroach acceptance cluster localcluster go github com cockroachdb cockroach acceptance testdocker go src github com cockroachdb cockroach acceptance util test go github com cockroachdb cockroach acceptance testdockersinglenode please assign take a look and update the issue accordingly
1
322,301
27,595,465,101
IssuesEvent
2023-03-09 05:50:46
antrea-io/antrea
https://api.github.com/repos/antrea-io/antrea
opened
windows e2e test used stale image on one Node
area/test area/test/e2e area/OS/windows kind/bug
**Describe the bug** <!-- A clear and concise description of what the bug is. If you believe this bug is a security issue, please don't use this template and follow our [security guidelines](/SECURITY.md) --> When I was debugging an issue on a windows testbed, I found one antrea-agent-windows Pod ran an old image. The correct version should be v1.11.0-dev-75a53913.dirty: ``` # antrea-agent-windows-nrpvd I0308 21:10:38.550971 6844 agent.go:99] Starting Antrea agent (version v1.11.0-dev-75a53913.dirty) # antrea-agent-windows-ztlsr I0308 21:09:47.755164 1484 agent.go:92] Starting Antrea agent (version v1.7.0-dev-2a37aec1.dirty) ``` Testbed deployment log: ``` [antrea-windows-e2e-proxyall-for-pull-request] $ /bin/sh -xe /tmp/jenkins8184209255165314312.sh + set -e + head -n1 ci/docker-registry + DOCKER_REGISTRY=projects.registry.vmware.com + sed -i /skipIfNotIPv4Cluster(t)/a\ \ \ \ \ \ \ \ skipIfHasWindowsNodes(t) test/e2e/nodeportlocal_test.go + ./ci/jenkins/test.sh --testcase windows-e2e --registry projects.registry.vmware.com --proxyall ===== Clean up stale files & folders older than 7 days under /tmp ===== ====== Cleanup Antrea Installation ====== No resources found in antrea-test namespace. ====== Building Antrea for the Following Commit ====== commit 75a5391315ac5d89f2a1e57f3cecbca3f816ade5 Merge: 32485497 18ddb089 Author: Kumar Atish <atish.iaf@gmail.com> Date: Wed Mar 8 23:03:12 2023 +0530 Merge 18ddb08974afc6d051c489fb87890a495841cb56 into 324854979d383c11e83e49b5c7e88d1caec6ebdd 7 1 pkg/agent/apiserver/handlers/memberlist/handler.go 62 26 pkg/agent/apiserver/handlers/memberlist/handler_test.go 5 1 pkg/support/dump.go Deleted Images: untagged: antrea/cni-binaries:v1.1.1 untagged: projects.registry.vmware.com/antrea/cni-binaries:v1.1.1 untagged: projects.registry.vmware.com/antrea/cni-binaries@sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219 deleted: sha256:bfaa431b732dd3d12776c98d0e684f65e9cf77c481b9e35496b3e1b02e25da05 deleted: sha256:fe0b72db62131760671074f192f8c182a0c22c6fa2292316123ebdf71ddb77f5 deleted: sha256:55139abd5e2a8db4f73283a7f50ed7b3b1d81d4d2956ec119a3d1f03c2eaeb4d deleted: sha256:7e7fc777a83c92df7fb369f3b2a61189218f567944bbf9269f69453372c109d9 deleted: sha256:25bb6906eea3f717d8295b99a4d69538624cedafa62c04bf7dff0b4ed97214d3 deleted: sha256:f6d6f546f40138b7b433a17e74af250fa533cbd34e24783d0ceef0a1b7fc7b64 deleted: sha256:67ce281807736e142e8151f56dd4a2e387d9389449c170eb7405c04761479ed4 untagged: ubuntu:22.04 untagged: projects.registry.vmware.com/antrea/ubuntu:22.04 untagged: projects.registry.vmware.com/antrea/ubuntu@sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b untagged: e2eteam/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1 deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4 deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9 deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3 deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88 deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463 deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68 deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac untagged: antrea/openvswitch-debs:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch-debs@sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668 deleted: sha256:fd116dd5d80d4f166894a08e1832afaf52fbbd6367dad185199094c996f62301 deleted: sha256:81f789fce63bda6d07e049856d386c61a636528699c76d022708ddee445b4b2e deleted: sha256:6b8158b5883be250d6f351957fb43c33a4e4c4b63898aa7a2a86522e0f0dc523 deleted: sha256:2eb4d98e07106613c35d3a505adcf490871ceeb3de9ce520c3b5ed1ab757ebf1 deleted: sha256:c960d4cb5b1e6437b012bf69ee311156c1b63ce80dae9bec4277b26b9947b742 deleted: sha256:e2b36c6bf533a216accfb28a94c4acb21ee631cab9b8eab1d723e65ce176d71a deleted: sha256:a68b22d0c835019f3f34a2eeabf873611708513b616483c091d94673ed9cc329 untagged: nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101 deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478 deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282 deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81 untagged: antrea/openvswitch:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch@sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83 untagged: antrea/base-ubuntu:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/base-ubuntu@sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d Total reclaimed space: 773.4MB BUILD_TAG: antrea-v1.11 22.04: Pulling from antrea/ubuntu Digest: sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b Status: Downloaded newer image for projects.registry.vmware.com/antrea/ubuntu:22.04 projects.registry.vmware.com/antrea/ubuntu:22.04 1.19: Pulling from antrea/golang Digest: sha256:6687844637736cc26e4a10386abb7b482e5d231574be0bef9260a255da1f4775 Status: Image is up to date for projects.registry.vmware.com/antrea/golang:1.19 projects.registry.vmware.com/antrea/golang:1.19 antrea-v1.11: Pulling from antrea/openvswitch-debs 76769433fd8a: Already exists 57d34a411582: Pulling fs layer 7dad3cb1dcc7: Pulling fs layer 4f21db537580: Pulling fs layer 7dad3cb1dcc7: Verifying Checksum 7dad3cb1dcc7: Download complete 4f21db537580: Verifying Checksum 4f21db537580: Download complete 57d34a411582: Verifying Checksum 57d34a411582: Download complete 57d34a411582: Pull complete 7dad3cb1dcc7: Pull complete 4f21db537580: Pull complete Digest: sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668 Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11 projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11 antrea-v1.11: Pulling from antrea/openvswitch Digest: sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83 Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11 projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11 v1.1.1: Pulling from antrea/cni-binaries 76769433fd8a: Already exists c96ff6af3775: Pulling fs layer 247dd6786f33: Pulling fs layer 247dd6786f33: Verifying Checksum 247dd6786f33: Download complete c96ff6af3775: Verifying Checksum c96ff6af3775: Download complete c96ff6af3775: Pull complete 247dd6786f33: Pull complete Digest: sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219 Status: Downloaded newer image for projects.registry.vmware.com/antrea/cni-binaries:v1.1.1 projects.registry.vmware.com/antrea/cni-binaries:v1.1.1 antrea-v1.11: Pulling from antrea/base-ubuntu Digest: sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d Status: Downloaded newer image for projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11 projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11 Sending build context to Docker daemon 24.06kB Step 1/5 : FROM ubuntu:22.04 as ovs-debs ---> 74f2314a03de Step 2/5 : ARG OVS_VERSION ---> Using cache ---> a68b22d0c835 Step 3/5 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev ---> Using cache ---> c960d4cb5b1e Step 4/5 : COPY apply-patches.sh / ---> Using cache ---> 6b8158b5883b Step 5/5 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch* ---> Using cache ---> fd116dd5d80d Successfully built fd116dd5d80d Successfully tagged antrea/openvswitch-debs:antrea-v1.11 Sending build context to Docker daemon 24.06kB Step 1/11 : FROM ubuntu:22.04 as ovs-debs ---> 74f2314a03de Step 2/11 : ARG OVS_VERSION ---> Using cache ---> a68b22d0c835 Step 3/11 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev ---> Using cache ---> c960d4cb5b1e Step 4/11 : COPY apply-patches.sh / ---> Using cache ---> 6b8158b5883b Step 5/11 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch* ---> Using cache ---> fd116dd5d80d Step 6/11 : FROM ubuntu:22.04 ---> 74f2314a03de Step 7/11 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Using cache ---> 2873291cca14 Step 8/11 : LABEL description="A Docker image based on Ubuntu 22.04 which includes Open vSwitch built from source." ---> Using cache ---> 0e26ec9b7d9d Step 9/11 : COPY --from=ovs-debs /tmp/ovs-debs/* /tmp/ovs-debs/ ---> Using cache ---> 0c4db4416cce Step 10/11 : COPY charon-logging.conf /tmp ---> Using cache ---> 79d23217a023 Step 11/11 : RUN apt-get update && apt-get install -y --no-install-recommends iptables logrotate libstrongswan-standard-plugins && (dpkg -i /tmp/ovs-debs/*.deb || apt-get -f -y --no-install-recommends install) && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && sed -i "/rotate /a\ #size 100M" /etc/logrotate.d/openvswitch-switch && sed -i "/^.*filelog.*{/r /tmp/charon-logging.conf" /etc/strongswan.d/charon-logging.conf && rm -rf /tmp/* ---> Using cache ---> 3852ac294105 Successfully built 3852ac294105 Successfully tagged antrea/openvswitch:antrea-v1.11 /var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request Sending build context to Docker daemon 13.31kB Step 1/7 : ARG BUILD_TAG Step 2/7 : FROM ubuntu:22.04 as cni-binaries ---> 74f2314a03de Step 3/7 : ARG CNI_BINARIES_VERSION ---> Using cache ---> 67ce28180773 Step 4/7 : ARG WHEREABOUTS_VERSION=v0.5.4 ---> Using cache ---> f6d6f546f401 Step 5/7 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates ---> Using cache ---> 7e7fc777a83c Step 6/7 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth" ---> Using cache ---> 55139abd5e2a Step 7/7 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts ---> Using cache ---> bfaa431b732d Successfully built bfaa431b732d Successfully tagged antrea/cni-binaries:v1.1.1 Sending build context to Docker daemon 13.31kB Step 1/15 : ARG BUILD_TAG Step 2/15 : FROM ubuntu:22.04 as cni-binaries ---> 74f2314a03de Step 3/15 : ARG CNI_BINARIES_VERSION ---> Using cache ---> 67ce28180773 Step 4/15 : ARG WHEREABOUTS_VERSION=v0.5.4 ---> Using cache ---> f6d6f546f401 Step 5/15 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates ---> Using cache ---> 7e7fc777a83c Step 6/15 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth" ---> Using cache ---> 55139abd5e2a Step 7/15 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts ---> Using cache ---> bfaa431b732d Step 8/15 : FROM antrea/openvswitch:${BUILD_TAG} ---> 3852ac294105 Step 9/15 : ARG SURICATA_VERSION ---> Using cache ---> bafddc6af04e Step 10/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Using cache ---> ec89fc76d92d Step 11/15 : LABEL description="An Ubuntu based Docker base image for Antrea." ---> Using cache ---> dd8fb640925d Step 12/15 : USER root ---> Using cache ---> 5c814d44d69b Step 13/15 : ADD https://raw.githubusercontent.com/kubernetes-sigs/iptables-wrappers/9e6ce59c864623ea71a6f7d59c35fcb13a919b87/iptables-wrapper-installer.sh /iptables-wrapper-installer.sh ---> Using cache ---> 789847397959 Step 14/15 : RUN apt-get update && apt-get install -y --no-install-recommends ipset jq inotify-tools gpg-agent software-properties-common && add-apt-repository ppa:oisf/suricata-${SURICATA_VERSION} && apt-get update && apt-get install -y suricata && apt-get remove -y gpg-agent software-properties-common && apt-get autoremove -y && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && chmod +x /iptables-wrapper-installer.sh && /iptables-wrapper-installer.sh ---> Using cache ---> 815257da934e Step 15/15 : COPY --from=cni-binaries /opt/cni/bin /opt/cni/bin ---> Using cache ---> ad75b5563247 Successfully built ad75b5563247 Successfully tagged antrea/base-ubuntu:antrea-v1.11 /var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request ===> Building Antrea bins and antrea/antrea-ubuntu Docker image <=== docker build -t antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.ubuntu --build-arg OVS_VERSION=2.17.3 --build-arg GO_VERSION=1.19 --build-arg BUILD_TAG=antrea-v1.11 . Sending build context to Docker daemon 47.41MB Step 1/15 : ARG GO_VERSION Step 2/15 : ARG BUILD_TAG Step 3/15 : FROM golang:${GO_VERSION} as antrea-build ---> ff3cd58379d9 Step 4/15 : WORKDIR /antrea ---> Running in 62533a8c8585 Removing intermediate container 62533a8c8585 ---> 869906fadac2 Step 5/15 : COPY go.mod /antrea/go.mod ---> e087b2973e1c Step 6/15 : RUN go mod download ---> Running in 6771daf43c3b Removing intermediate container 6771daf43c3b ---> 41125c6f24ee Step 7/15 : COPY . /antrea ---> de288ca18a4c Step 8/15 : RUN make antrea-agent antrea-controller antrea-cni antctl-linux ---> Running in a3938237789d GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-agent GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-controller GOOS=linux CGO_ENABLED=0 go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni Removing intermediate container a3938237789d ---> 848dec0d7aaf Step 9/15 : RUN mv bin/antctl-linux bin/antctl ---> Running in b5313b457af3 Removing intermediate container b5313b457af3 ---> f6bfc1318921 Step 10/15 : FROM antrea/base-ubuntu:${BUILD_TAG} ---> ad75b5563247 Step 11/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Using cache ---> 54aa6d3d57c4 Step 12/15 : LABEL description="The Docker image to deploy the Antrea CNI." ---> Using cache ---> 8161ff16a1cb Step 13/15 : USER root ---> Using cache ---> ef8b9278df6d Step 14/15 : COPY build/images/scripts/* /usr/local/bin/ ---> Using cache ---> afd46dfce8c0 Step 15/15 : COPY --from=antrea-build /antrea/bin/* /usr/local/bin/ ---> 2d3d7425fc7f [Warning] One or more build-args [OVS_VERSION] were not consumed Successfully built 2d3d7425fc7f Successfully tagged antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty docker tag antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty antrea/antrea-ubuntu ====== Delivering Antrea to all Nodes ====== ====== Updating yaml files to enable proxyAll ====== ===== Pull necessary images on Control-Plane node ===== projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/nginx:1.15-alpine ===== Deliver Antrea to Linux worker nodes and pull necessary images on worker nodes ===== sending incremental file list antrea-ubuntu.tar 45,216 0% 0.00kB/s 0:00:00 158,854,896 31% 151.45MB/s 0:00:02 212,998,875 41% 101.60MB/s 0:00:02 282,112,496 55% 89.73MB/s 0:00:02 349,690,560 68% 83.42MB/s 0:00:01 402,000,296 78% 58.01MB/s 0:00:01 461,453,972 90% 59.28MB/s 0:00:00 511,165,952 100% 72.18MB/s 0:00:06 (xfr#1, to-chk=0/1) sent 287,388,834 bytes received 158,320 bytes 33,829,076.94 bytes/sec total size is 511,165,952 speedup is 1.78 Deleted Images: untagged: k8s.gcr.io/e2e-test-images/agnhost:2.29 untagged: k8s.gcr.io/e2e-test-images/agnhost@sha256:1ae25df1a1985fa815f712b051f467dcd2d016d9efb74ca41ef385f9ed64ada9 deleted: sha256:3f4d3ab351ab4180a2c0fc655ba97263ba74fa353d41642b3e09bc16c231b735 deleted: sha256:9735e33cc1701e722d4ef42086090078232f4d45087ebea8a40de28b18eb190d deleted: sha256:10f54c8640b3c21f6a7575ed67fa29c1811ed0a1404b0d398ab197ef2a55c46b deleted: sha256:4104722c5b5696347fb190e2e341ac593a0dd3953ad63bf82c89e037f71da341 deleted: sha256:7abc3633883948ff206528cd56bada29a18222fd642bee4c76d67ab12e590113 deleted: sha256:41758cdb2ea620f0c2db17598ce779841c9d4dce6cb6c457b302e082b2d7771f deleted: sha256:60e550c19b857d2261358edfa8b11cc43e53a076f3baea759553ce914889bd0d deleted: sha256:aa75a9d85b48eba258a28d6975fb3db0393809cc290d3dab4d0bf2bdb97d1e79 deleted: sha256:891c281852b0ab60d0a307ea6c5fbd0697004ade3745e7adcfdd549391dd633f deleted: sha256:33e8713114f88c8cb3f60c8a0a4aefe2500823b2fbbae05488b0185ba226caae untagged: e2eteam/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1 deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4 deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9 deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3 deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88 deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463 deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68 deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac untagged: nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101 deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478 deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282 deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81 Total reclaimed space: 253.1MB The image antrea/antrea-ubuntu:latest already exists, renaming the old one with ID sha256:568e9f48739ea51ab066586797f29459d1710bf9608f8ab7acf591d99b77bf73 to empty string Loaded image: antrea/antrea-ubuntu:latest projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/nginx:1.15-alpine k8s.gcr.io/e2e-test-images/agnhost:2.29 ===== Deliver Antrea Windows to Windows worker nodes and pull necessary images on Windows worker nodes ===== ==== Reverting Windows VM a-ms-0008-win-0 ===== Windows VM a-ms-0008-win-0 powered on projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.29 projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0 projects.registry.vmware.com/antrea/e2eteam-pause:3.2 Error response from daemon: Get https://mcr.microsoft.com/v2/: x509: certificate has expired or is not yet valid === Build Windows on Windows Node=== 1.19-nanoserver: Pulling from antrea/golang 5c9d6483dab1: Pulling fs layer 9982991b8208: Pulling fs layer cb374f741579: Pulling fs layer 732dbfa4c42c: Pulling fs layer 64bbe06a1892: Pulling fs layer 15263a6e39d7: Pulling fs layer d1ebeb495bd3: Pulling fs layer 460984e65c46: Pulling fs layer 109a63e20d99: Pulling fs layer e3c52c6cbb5c: Pulling fs layer 64bbe06a1892: Waiting 15263a6e39d7: Waiting d1ebeb495bd3: Waiting 460984e65c46: Waiting 109a63e20d99: Waiting e3c52c6cbb5c: Waiting 732dbfa4c42c: Waiting 9982991b8208: Verifying Checksum 9982991b8208: Download complete cb374f741579: Verifying Checksum cb374f741579: Download complete 732dbfa4c42c: Verifying Checksum 732dbfa4c42c: Download complete 64bbe06a1892: Verifying Checksum 64bbe06a1892: Download complete 15263a6e39d7: Verifying Checksum 15263a6e39d7: Download complete d1ebeb495bd3: Verifying Checksum d1ebeb495bd3: Download complete 109a63e20d99: Download complete e3c52c6cbb5c: Verifying Checksum e3c52c6cbb5c: Download complete 5c9d6483dab1: Verifying Checksum 5c9d6483dab1: Download complete 460984e65c46: Verifying Checksum 460984e65c46: Download complete 5c9d6483dab1: Pull complete 9982991b8208: Pull complete cb374f741579: Pull complete 732dbfa4c42c: Pull complete 64bbe06a1892: Pull complete 15263a6e39d7: Pull complete d1ebeb495bd3: Pull complete 460984e65c46: Pull complete 109a63e20d99: Pull complete e3c52c6cbb5c: Pull complete Digest: sha256:d5843e74790da22bd3d84d098dcbe21c7318a7c7c4db6bf1de09cb193e64c513 Status: Downloaded newer image for projects.registry.vmware.com/antrea/golang:1.19-nanoserver projects.registry.vmware.com/antrea/golang:1.19-nanoserver ===> Building Antrea bins and antrea/antrea-windows Docker image <=== docker build --pull -t antrea/antrea-windows:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.windows --network host --build-arg GO_VERSION=1.19 --build-arg CNI_BINARIES_VERSION=v1.1.1 --build-arg NANOSERVER_VERSION=1809 --build-arg WIN_BUILD_TAG=caa8cac3d3 . Sending build context to Docker daemon 558.6MB Step 1/21 : ARG WIN_BUILD_TAG Step 2/21 : ARG NANOSERVER_VERSION Step 3/21 : FROM antrea/base-windows:${WIN_BUILD_TAG} as antrea-build-windows caa8cac3d3: Pulling from antrea/base-windows 5ead999142ec: Pulling fs layer 3724c0f91ae7: Pulling fs layer 97a8c2800739: Pulling fs layer c038a2fdb0b5: Pulling fs layer 5ef04caec17a: Pulling fs layer 8bca1755b69d: Pulling fs layer 3ebc2faa6c3e: Pulling fs layer d85edb0b9152: Pulling fs layer b343c39a6010: Pulling fs layer b673c002e661: Pulling fs layer 99b72f4231b5: Pulling fs layer d85edb0b9152: Waiting 3ebc2faa6c3e: Waiting 8bca1755b69d: Waiting c038a2fdb0b5: Waiting b343c39a6010: Waiting 5ef04caec17a: Waiting b673c002e661: Waiting 99b72f4231b5: Waiting 97a8c2800739: Download complete 3724c0f91ae7: Download complete 5ead999142ec: Verifying Checksum 5ead999142ec: Download complete c038a2fdb0b5: Download complete 5ef04caec17a: Verifying Checksum 5ef04caec17a: Download complete 8bca1755b69d: Verifying Checksum 8bca1755b69d: Download complete b343c39a6010: Verifying Checksum b343c39a6010: Download complete 3ebc2faa6c3e: Verifying Checksum 3ebc2faa6c3e: Download complete 99b72f4231b5: Verifying Checksum 99b72f4231b5: Download complete b673c002e661: Verifying Checksum b673c002e661: Download complete d85edb0b9152: Verifying Checksum d85edb0b9152: Download complete 5ead999142ec: Pull complete 3724c0f91ae7: Pull complete 97a8c2800739: Pull complete c038a2fdb0b5: Pull complete 5ef04caec17a: Pull complete 8bca1755b69d: Pull complete 3ebc2faa6c3e: Pull complete d85edb0b9152: Pull complete b343c39a6010: Pull complete b673c002e661: Pull complete 99b72f4231b5: Pull complete Digest: sha256:c6edebf1a80553a5d474b6663657e69bc5b2cf54b116a1759e4d053c609a575d Status: Downloaded newer image for antrea/base-windows:caa8cac3d3 ---> 79d3453071f3 Step 4/21 : WORKDIR /antrea ---> Running in 8019bfcc22c9 Removing intermediate container 8019bfcc22c9 ---> 6f713bbf1027 Step 5/21 : COPY go.mod /antrea/go.mod ---> 30bbece0963c Step 6/21 : RUN go mod download ---> Running in 7a09a9c079f4 Removing intermediate container 7a09a9c079f4 ---> af9b4babd43e Step 7/21 : COPY . /antrea ---> 60d82c848851 Step 8/21 : RUN sh -c 'make windows-bin' ---> Running in aab89dd36ca9 GOOS=windows CGO_ENABLED=0 go build -o C:/antrea/bin -ldflags ' -X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni antrea.io/antrea/cmd/antrea-agent antrea.io/antrea/cmd/antctl Removing intermediate container aab89dd36ca9 ---> a06472a57304 Step 9/21 : FROM mcr.microsoft.com/powershell:lts-nanoserver-${NANOSERVER_VERSION} lts-nanoserver-1809: Pulling from powershell af0153d864f1: Pulling fs layer eb9aadfb9ebf: Pulling fs layer 698ae59174d5: Pulling fs layer ca1ef8c236fc: Pulling fs layer e3789ff6b796: Pulling fs layer a9a89a12f338: Pulling fs layer ccb30ac1f365: Pulling fs layer 0b2e07545fa8: Pulling fs layer 72408742a0d5: Pulling fs layer ca1ef8c236fc: Waiting 0b2e07545fa8: Waiting 72408742a0d5: Waiting e3789ff6b796: Waiting a9a89a12f338: Waiting ccb30ac1f365: Waiting eb9aadfb9ebf: Verifying Checksum eb9aadfb9ebf: Download complete 698ae59174d5: Verifying Checksum 698ae59174d5: Download complete ca1ef8c236fc: Verifying Checksum ca1ef8c236fc: Download complete e3789ff6b796: Verifying Checksum e3789ff6b796: Download complete af0153d864f1: Verifying Checksum af0153d864f1: Download complete ccb30ac1f365: Verifying Checksum ccb30ac1f365: Download complete 0b2e07545fa8: Verifying Checksum 0b2e07545fa8: Download complete 72408742a0d5: Verifying Checksum 72408742a0d5: Download complete a9a89a12f338: Verifying Checksum a9a89a12f338: Download complete af0153d864f1: Pull complete eb9aadfb9ebf: Pull complete 698ae59174d5: Pull complete ca1ef8c236fc: Pull complete e3789ff6b796: Pull complete a9a89a12f338: Pull complete ccb30ac1f365: Pull complete 0b2e07545fa8: Pull complete 72408742a0d5: Pull complete Digest: sha256:41643cdb0d064e3a282badd7c1568e8ffa904536a3e95116ce353696a5d1d471 Status: Downloaded newer image for mcr.microsoft.com/powershell:lts-nanoserver-1809 ---> d57d56a6e450 Step 10/21 : SHELL ["pwsh", "-NoLogo", "-Command", "$ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';"] ---> Running in 494b3abb7438 Removing intermediate container 494b3abb7438 ---> ad61f45fa627 Step 11/21 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Running in 3c3062b996af Removing intermediate container 3c3062b996af ---> 4362e1e85c5d Step 12/21 : LABEL description="A Docker image to deploy the Antrea CNI." ---> Running in 950d9303ee7c Removing intermediate container 950d9303ee7c ---> a6d107b46ad6 Step 13/21 : USER ContainerAdministrator ---> Running in 08026f87f27e Removing intermediate container 08026f87f27e ---> 5ed0dd7211b1 Step 14/21 : RUN mkdir -Force C:\k\antrea\bin ---> Running in 53264a32f2da Directory: C:\k\antrea Mode LastWriteTime Length Name ---- ------------- ------ ---- d---- 3/8/2023 8:54 PM bin Removing intermediate container 53264a32f2da ---> 744d452bdcef Step 15/21 : COPY --from=antrea-build-windows /opt/cni/bin /k/antrea/cni ---> da2f7aea09d2 Step 16/21 : COPY --from=antrea-build-windows /antrea/build/images/scripts/Install-WindowsCNI.ps1 /k/antrea/ ---> b3509d1238ca Step 17/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-agent.exe /k/antrea/bin/ ---> c75b95fbe770 Step 18/21 : COPY --from=antrea-build-windows /antrea/bin/antctl.exe /k/antrea/bin/antctl.exe ---> a6560ddef3ad Step 19/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-cni.exe /k/antrea/cni/antrea.exe ---> ddf94081e150 Step 20/21 : RUN mkdir C:\k\antrea\utils ---> Running in 5e156644b55e Directory: C:\k\antrea Mode LastWriteTime Length Name ---- ------------- ------ ---- d---- 3/8/2023 8:55 PM utils Removing intermediate container 5e156644b55e ---> db43d08befea Step 21/21 : COPY --from=antrea-build-windows /wins/wins.exe /k/antrea/utils/wins.exe ---> f84ef4d5fd9d [Warning] One or more build-args [CNI_BINARIES_VERSION GO_VERSION] were not consumed Successfully built f84ef4d5fd9d Successfully tagged antrea/antrea-windows:v1.11.0-dev-75a53913.dirty docker tag antrea/antrea-windows:v1.11.0-dev-75a53913.dirty antrea/antrea-windows ==== Reverting Windows VM a-ms-0008-win-1 ===== Windows VM a-ms-0008-win-1 powered on The command completed successfully. projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.29 projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0 projects.registry.vmware.com/antrea/e2eteam-pause:3.2 mcr.microsoft.com/windows/servercore/iis:latest The image projects.registry.vmware.com/antrea/antrea-windows:latest already exists, renaming the old one with ID sha256:4653103584af17d23287e7ce178d037e588f8188506402323dc0dff943264b2b to empty string Loaded image: projects.registry.vmware.com/antrea/antrea-windows:latest ====== Running Antrea e2e Tests ====== customresourcedefinition.apiextensions.k8s.io/antreaagentinfos.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/antreacontrollerinfos.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/clustergroups.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/clusternetworkpolicies.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/egresses.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/externalentities.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/externalippools.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/externalnodes.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/ippools.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/supportbundlecollections.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/tiers.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/traceflows.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/trafficcontrols.crd.antrea.io created serviceaccount/antrea-agent created serviceaccount/antctl created serviceaccount/antrea-controller created secret/antrea-agent-service-account-token created secret/antctl-service-account-token created configmap/antrea-config created customresourcedefinition.apiextensions.k8s.io/groups.crd.antrea.io created clusterrole.rbac.authorization.k8s.io/antrea-agent created clusterrole.rbac.authorization.k8s.io/antctl created clusterrole.rbac.authorization.k8s.io/antrea-cluster-identity-reader created clusterrole.rbac.authorization.k8s.io/antrea-controller created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-edit created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-view created clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-edit created clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-view created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-edit created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-view created clusterrolebinding.rbac.authorization.k8s.io/antrea-agent created clusterrolebinding.rbac.authorization.k8s.io/antctl created clusterrolebinding.rbac.authorization.k8s.io/antrea-controller created service/antrea created daemonset.apps/antrea-agent created deployment.apps/antrea-controller created apiservice.apiregistration.k8s.io/v1beta2.controlplane.antrea.io created apiservice.apiregistration.k8s.io/v1beta1.system.antrea.io created apiservice.apiregistration.k8s.io/v1alpha1.stats.antrea.io created mutatingwebhookconfiguration.admissionregistration.k8s.io/crdmutator.antrea.io created validatingwebhookconfiguration.admissionregistration.k8s.io/crdvalidator.antrea.io created configmap/antrea-agent-windows-kht6m7hthm created configmap/antrea-windows-config-db2774h9dt created daemonset.apps/antrea-agent-windows created deployment.apps/coredns restarted Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated... Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated... Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated... Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination... Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination... Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination... Waiting for deployment "coredns" rollout to finish: 1 of 2 updated replicas are available... deployment "coredns" successfully rolled out deployment "antrea-controller" successfully rolled out daemon set "antrea-agent" successfully rolled out Waiting for daemon set "antrea-agent-windows" rollout to finish: 0 of 2 updated pods are available... Waiting for daemon set "antrea-agent-windows" rollout to finish: 1 of 2 updated pods are available... daemon set "antrea-agent-windows" successfully rolled out Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-BF-B1 10 Gbps Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-B7-EE 10 Gbps === Generate ssh-config === Generating ssh-config for Node a-ms-0008-0 Generating ssh-config for Node a-ms-0008-1 Generating ssh-config for Node a-ms-0008-win-0 Generating ssh-config for Node a-ms-0008-win-1 ====== Run test with e2e test ====== ... ```
2.0
windows e2e test used stale image on one Node - **Describe the bug** <!-- A clear and concise description of what the bug is. If you believe this bug is a security issue, please don't use this template and follow our [security guidelines](/SECURITY.md) --> When I was debugging an issue on a windows testbed, I found one antrea-agent-windows Pod ran an old image. The correct version should be v1.11.0-dev-75a53913.dirty: ``` # antrea-agent-windows-nrpvd I0308 21:10:38.550971 6844 agent.go:99] Starting Antrea agent (version v1.11.0-dev-75a53913.dirty) # antrea-agent-windows-ztlsr I0308 21:09:47.755164 1484 agent.go:92] Starting Antrea agent (version v1.7.0-dev-2a37aec1.dirty) ``` Testbed deployment log: ``` [antrea-windows-e2e-proxyall-for-pull-request] $ /bin/sh -xe /tmp/jenkins8184209255165314312.sh + set -e + head -n1 ci/docker-registry + DOCKER_REGISTRY=projects.registry.vmware.com + sed -i /skipIfNotIPv4Cluster(t)/a\ \ \ \ \ \ \ \ skipIfHasWindowsNodes(t) test/e2e/nodeportlocal_test.go + ./ci/jenkins/test.sh --testcase windows-e2e --registry projects.registry.vmware.com --proxyall ===== Clean up stale files & folders older than 7 days under /tmp ===== ====== Cleanup Antrea Installation ====== No resources found in antrea-test namespace. ====== Building Antrea for the Following Commit ====== commit 75a5391315ac5d89f2a1e57f3cecbca3f816ade5 Merge: 32485497 18ddb089 Author: Kumar Atish <atish.iaf@gmail.com> Date: Wed Mar 8 23:03:12 2023 +0530 Merge 18ddb08974afc6d051c489fb87890a495841cb56 into 324854979d383c11e83e49b5c7e88d1caec6ebdd 7 1 pkg/agent/apiserver/handlers/memberlist/handler.go 62 26 pkg/agent/apiserver/handlers/memberlist/handler_test.go 5 1 pkg/support/dump.go Deleted Images: untagged: antrea/cni-binaries:v1.1.1 untagged: projects.registry.vmware.com/antrea/cni-binaries:v1.1.1 untagged: projects.registry.vmware.com/antrea/cni-binaries@sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219 deleted: sha256:bfaa431b732dd3d12776c98d0e684f65e9cf77c481b9e35496b3e1b02e25da05 deleted: sha256:fe0b72db62131760671074f192f8c182a0c22c6fa2292316123ebdf71ddb77f5 deleted: sha256:55139abd5e2a8db4f73283a7f50ed7b3b1d81d4d2956ec119a3d1f03c2eaeb4d deleted: sha256:7e7fc777a83c92df7fb369f3b2a61189218f567944bbf9269f69453372c109d9 deleted: sha256:25bb6906eea3f717d8295b99a4d69538624cedafa62c04bf7dff0b4ed97214d3 deleted: sha256:f6d6f546f40138b7b433a17e74af250fa533cbd34e24783d0ceef0a1b7fc7b64 deleted: sha256:67ce281807736e142e8151f56dd4a2e387d9389449c170eb7405c04761479ed4 untagged: ubuntu:22.04 untagged: projects.registry.vmware.com/antrea/ubuntu:22.04 untagged: projects.registry.vmware.com/antrea/ubuntu@sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b untagged: e2eteam/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1 deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4 deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9 deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3 deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88 deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463 deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68 deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac untagged: antrea/openvswitch-debs:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch-debs@sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668 deleted: sha256:fd116dd5d80d4f166894a08e1832afaf52fbbd6367dad185199094c996f62301 deleted: sha256:81f789fce63bda6d07e049856d386c61a636528699c76d022708ddee445b4b2e deleted: sha256:6b8158b5883be250d6f351957fb43c33a4e4c4b63898aa7a2a86522e0f0dc523 deleted: sha256:2eb4d98e07106613c35d3a505adcf490871ceeb3de9ce520c3b5ed1ab757ebf1 deleted: sha256:c960d4cb5b1e6437b012bf69ee311156c1b63ce80dae9bec4277b26b9947b742 deleted: sha256:e2b36c6bf533a216accfb28a94c4acb21ee631cab9b8eab1d723e65ce176d71a deleted: sha256:a68b22d0c835019f3f34a2eeabf873611708513b616483c091d94673ed9cc329 untagged: nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101 deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478 deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282 deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81 untagged: antrea/openvswitch:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/openvswitch@sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83 untagged: antrea/base-ubuntu:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11 untagged: projects.registry.vmware.com/antrea/base-ubuntu@sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d Total reclaimed space: 773.4MB BUILD_TAG: antrea-v1.11 22.04: Pulling from antrea/ubuntu Digest: sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b Status: Downloaded newer image for projects.registry.vmware.com/antrea/ubuntu:22.04 projects.registry.vmware.com/antrea/ubuntu:22.04 1.19: Pulling from antrea/golang Digest: sha256:6687844637736cc26e4a10386abb7b482e5d231574be0bef9260a255da1f4775 Status: Image is up to date for projects.registry.vmware.com/antrea/golang:1.19 projects.registry.vmware.com/antrea/golang:1.19 antrea-v1.11: Pulling from antrea/openvswitch-debs 76769433fd8a: Already exists 57d34a411582: Pulling fs layer 7dad3cb1dcc7: Pulling fs layer 4f21db537580: Pulling fs layer 7dad3cb1dcc7: Verifying Checksum 7dad3cb1dcc7: Download complete 4f21db537580: Verifying Checksum 4f21db537580: Download complete 57d34a411582: Verifying Checksum 57d34a411582: Download complete 57d34a411582: Pull complete 7dad3cb1dcc7: Pull complete 4f21db537580: Pull complete Digest: sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668 Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11 projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11 antrea-v1.11: Pulling from antrea/openvswitch Digest: sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83 Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11 projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11 v1.1.1: Pulling from antrea/cni-binaries 76769433fd8a: Already exists c96ff6af3775: Pulling fs layer 247dd6786f33: Pulling fs layer 247dd6786f33: Verifying Checksum 247dd6786f33: Download complete c96ff6af3775: Verifying Checksum c96ff6af3775: Download complete c96ff6af3775: Pull complete 247dd6786f33: Pull complete Digest: sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219 Status: Downloaded newer image for projects.registry.vmware.com/antrea/cni-binaries:v1.1.1 projects.registry.vmware.com/antrea/cni-binaries:v1.1.1 antrea-v1.11: Pulling from antrea/base-ubuntu Digest: sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d Status: Downloaded newer image for projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11 projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11 Sending build context to Docker daemon 24.06kB Step 1/5 : FROM ubuntu:22.04 as ovs-debs ---> 74f2314a03de Step 2/5 : ARG OVS_VERSION ---> Using cache ---> a68b22d0c835 Step 3/5 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev ---> Using cache ---> c960d4cb5b1e Step 4/5 : COPY apply-patches.sh / ---> Using cache ---> 6b8158b5883b Step 5/5 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch* ---> Using cache ---> fd116dd5d80d Successfully built fd116dd5d80d Successfully tagged antrea/openvswitch-debs:antrea-v1.11 Sending build context to Docker daemon 24.06kB Step 1/11 : FROM ubuntu:22.04 as ovs-debs ---> 74f2314a03de Step 2/11 : ARG OVS_VERSION ---> Using cache ---> a68b22d0c835 Step 3/11 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev ---> Using cache ---> c960d4cb5b1e Step 4/11 : COPY apply-patches.sh / ---> Using cache ---> 6b8158b5883b Step 5/11 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch* ---> Using cache ---> fd116dd5d80d Step 6/11 : FROM ubuntu:22.04 ---> 74f2314a03de Step 7/11 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Using cache ---> 2873291cca14 Step 8/11 : LABEL description="A Docker image based on Ubuntu 22.04 which includes Open vSwitch built from source." ---> Using cache ---> 0e26ec9b7d9d Step 9/11 : COPY --from=ovs-debs /tmp/ovs-debs/* /tmp/ovs-debs/ ---> Using cache ---> 0c4db4416cce Step 10/11 : COPY charon-logging.conf /tmp ---> Using cache ---> 79d23217a023 Step 11/11 : RUN apt-get update && apt-get install -y --no-install-recommends iptables logrotate libstrongswan-standard-plugins && (dpkg -i /tmp/ovs-debs/*.deb || apt-get -f -y --no-install-recommends install) && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && sed -i "/rotate /a\ #size 100M" /etc/logrotate.d/openvswitch-switch && sed -i "/^.*filelog.*{/r /tmp/charon-logging.conf" /etc/strongswan.d/charon-logging.conf && rm -rf /tmp/* ---> Using cache ---> 3852ac294105 Successfully built 3852ac294105 Successfully tagged antrea/openvswitch:antrea-v1.11 /var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request Sending build context to Docker daemon 13.31kB Step 1/7 : ARG BUILD_TAG Step 2/7 : FROM ubuntu:22.04 as cni-binaries ---> 74f2314a03de Step 3/7 : ARG CNI_BINARIES_VERSION ---> Using cache ---> 67ce28180773 Step 4/7 : ARG WHEREABOUTS_VERSION=v0.5.4 ---> Using cache ---> f6d6f546f401 Step 5/7 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates ---> Using cache ---> 7e7fc777a83c Step 6/7 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth" ---> Using cache ---> 55139abd5e2a Step 7/7 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts ---> Using cache ---> bfaa431b732d Successfully built bfaa431b732d Successfully tagged antrea/cni-binaries:v1.1.1 Sending build context to Docker daemon 13.31kB Step 1/15 : ARG BUILD_TAG Step 2/15 : FROM ubuntu:22.04 as cni-binaries ---> 74f2314a03de Step 3/15 : ARG CNI_BINARIES_VERSION ---> Using cache ---> 67ce28180773 Step 4/15 : ARG WHEREABOUTS_VERSION=v0.5.4 ---> Using cache ---> f6d6f546f401 Step 5/15 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates ---> Using cache ---> 7e7fc777a83c Step 6/15 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth" ---> Using cache ---> 55139abd5e2a Step 7/15 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts ---> Using cache ---> bfaa431b732d Step 8/15 : FROM antrea/openvswitch:${BUILD_TAG} ---> 3852ac294105 Step 9/15 : ARG SURICATA_VERSION ---> Using cache ---> bafddc6af04e Step 10/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Using cache ---> ec89fc76d92d Step 11/15 : LABEL description="An Ubuntu based Docker base image for Antrea." ---> Using cache ---> dd8fb640925d Step 12/15 : USER root ---> Using cache ---> 5c814d44d69b Step 13/15 : ADD https://raw.githubusercontent.com/kubernetes-sigs/iptables-wrappers/9e6ce59c864623ea71a6f7d59c35fcb13a919b87/iptables-wrapper-installer.sh /iptables-wrapper-installer.sh ---> Using cache ---> 789847397959 Step 14/15 : RUN apt-get update && apt-get install -y --no-install-recommends ipset jq inotify-tools gpg-agent software-properties-common && add-apt-repository ppa:oisf/suricata-${SURICATA_VERSION} && apt-get update && apt-get install -y suricata && apt-get remove -y gpg-agent software-properties-common && apt-get autoremove -y && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && chmod +x /iptables-wrapper-installer.sh && /iptables-wrapper-installer.sh ---> Using cache ---> 815257da934e Step 15/15 : COPY --from=cni-binaries /opt/cni/bin /opt/cni/bin ---> Using cache ---> ad75b5563247 Successfully built ad75b5563247 Successfully tagged antrea/base-ubuntu:antrea-v1.11 /var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request ===> Building Antrea bins and antrea/antrea-ubuntu Docker image <=== docker build -t antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.ubuntu --build-arg OVS_VERSION=2.17.3 --build-arg GO_VERSION=1.19 --build-arg BUILD_TAG=antrea-v1.11 . Sending build context to Docker daemon 47.41MB Step 1/15 : ARG GO_VERSION Step 2/15 : ARG BUILD_TAG Step 3/15 : FROM golang:${GO_VERSION} as antrea-build ---> ff3cd58379d9 Step 4/15 : WORKDIR /antrea ---> Running in 62533a8c8585 Removing intermediate container 62533a8c8585 ---> 869906fadac2 Step 5/15 : COPY go.mod /antrea/go.mod ---> e087b2973e1c Step 6/15 : RUN go mod download ---> Running in 6771daf43c3b Removing intermediate container 6771daf43c3b ---> 41125c6f24ee Step 7/15 : COPY . /antrea ---> de288ca18a4c Step 8/15 : RUN make antrea-agent antrea-controller antrea-cni antctl-linux ---> Running in a3938237789d GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-agent GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-controller GOOS=linux CGO_ENABLED=0 go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni Removing intermediate container a3938237789d ---> 848dec0d7aaf Step 9/15 : RUN mv bin/antctl-linux bin/antctl ---> Running in b5313b457af3 Removing intermediate container b5313b457af3 ---> f6bfc1318921 Step 10/15 : FROM antrea/base-ubuntu:${BUILD_TAG} ---> ad75b5563247 Step 11/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Using cache ---> 54aa6d3d57c4 Step 12/15 : LABEL description="The Docker image to deploy the Antrea CNI." ---> Using cache ---> 8161ff16a1cb Step 13/15 : USER root ---> Using cache ---> ef8b9278df6d Step 14/15 : COPY build/images/scripts/* /usr/local/bin/ ---> Using cache ---> afd46dfce8c0 Step 15/15 : COPY --from=antrea-build /antrea/bin/* /usr/local/bin/ ---> 2d3d7425fc7f [Warning] One or more build-args [OVS_VERSION] were not consumed Successfully built 2d3d7425fc7f Successfully tagged antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty docker tag antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty antrea/antrea-ubuntu ====== Delivering Antrea to all Nodes ====== ====== Updating yaml files to enable proxyAll ====== ===== Pull necessary images on Control-Plane node ===== projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/nginx:1.15-alpine ===== Deliver Antrea to Linux worker nodes and pull necessary images on worker nodes ===== sending incremental file list antrea-ubuntu.tar 45,216 0% 0.00kB/s 0:00:00 158,854,896 31% 151.45MB/s 0:00:02 212,998,875 41% 101.60MB/s 0:00:02 282,112,496 55% 89.73MB/s 0:00:02 349,690,560 68% 83.42MB/s 0:00:01 402,000,296 78% 58.01MB/s 0:00:01 461,453,972 90% 59.28MB/s 0:00:00 511,165,952 100% 72.18MB/s 0:00:06 (xfr#1, to-chk=0/1) sent 287,388,834 bytes received 158,320 bytes 33,829,076.94 bytes/sec total size is 511,165,952 speedup is 1.78 Deleted Images: untagged: k8s.gcr.io/e2e-test-images/agnhost:2.29 untagged: k8s.gcr.io/e2e-test-images/agnhost@sha256:1ae25df1a1985fa815f712b051f467dcd2d016d9efb74ca41ef385f9ed64ada9 deleted: sha256:3f4d3ab351ab4180a2c0fc655ba97263ba74fa353d41642b3e09bc16c231b735 deleted: sha256:9735e33cc1701e722d4ef42086090078232f4d45087ebea8a40de28b18eb190d deleted: sha256:10f54c8640b3c21f6a7575ed67fa29c1811ed0a1404b0d398ab197ef2a55c46b deleted: sha256:4104722c5b5696347fb190e2e341ac593a0dd3953ad63bf82c89e037f71da341 deleted: sha256:7abc3633883948ff206528cd56bada29a18222fd642bee4c76d67ab12e590113 deleted: sha256:41758cdb2ea620f0c2db17598ce779841c9d4dce6cb6c457b302e082b2d7771f deleted: sha256:60e550c19b857d2261358edfa8b11cc43e53a076f3baea759553ce914889bd0d deleted: sha256:aa75a9d85b48eba258a28d6975fb3db0393809cc290d3dab4d0bf2bdb97d1e79 deleted: sha256:891c281852b0ab60d0a307ea6c5fbd0697004ade3745e7adcfdd549391dd633f deleted: sha256:33e8713114f88c8cb3f60c8a0a4aefe2500823b2fbbae05488b0185ba226caae untagged: e2eteam/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost:2.13 untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1 deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4 deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9 deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3 deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88 deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463 deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68 deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac untagged: nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101 deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478 deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282 deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81 Total reclaimed space: 253.1MB The image antrea/antrea-ubuntu:latest already exists, renaming the old one with ID sha256:568e9f48739ea51ab066586797f29459d1710bf9608f8ab7acf591d99b77bf73 to empty string Loaded image: antrea/antrea-ubuntu:latest projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/nginx:1.15-alpine k8s.gcr.io/e2e-test-images/agnhost:2.29 ===== Deliver Antrea Windows to Windows worker nodes and pull necessary images on Windows worker nodes ===== ==== Reverting Windows VM a-ms-0008-win-0 ===== Windows VM a-ms-0008-win-0 powered on projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.29 projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0 projects.registry.vmware.com/antrea/e2eteam-pause:3.2 Error response from daemon: Get https://mcr.microsoft.com/v2/: x509: certificate has expired or is not yet valid === Build Windows on Windows Node=== 1.19-nanoserver: Pulling from antrea/golang 5c9d6483dab1: Pulling fs layer 9982991b8208: Pulling fs layer cb374f741579: Pulling fs layer 732dbfa4c42c: Pulling fs layer 64bbe06a1892: Pulling fs layer 15263a6e39d7: Pulling fs layer d1ebeb495bd3: Pulling fs layer 460984e65c46: Pulling fs layer 109a63e20d99: Pulling fs layer e3c52c6cbb5c: Pulling fs layer 64bbe06a1892: Waiting 15263a6e39d7: Waiting d1ebeb495bd3: Waiting 460984e65c46: Waiting 109a63e20d99: Waiting e3c52c6cbb5c: Waiting 732dbfa4c42c: Waiting 9982991b8208: Verifying Checksum 9982991b8208: Download complete cb374f741579: Verifying Checksum cb374f741579: Download complete 732dbfa4c42c: Verifying Checksum 732dbfa4c42c: Download complete 64bbe06a1892: Verifying Checksum 64bbe06a1892: Download complete 15263a6e39d7: Verifying Checksum 15263a6e39d7: Download complete d1ebeb495bd3: Verifying Checksum d1ebeb495bd3: Download complete 109a63e20d99: Download complete e3c52c6cbb5c: Verifying Checksum e3c52c6cbb5c: Download complete 5c9d6483dab1: Verifying Checksum 5c9d6483dab1: Download complete 460984e65c46: Verifying Checksum 460984e65c46: Download complete 5c9d6483dab1: Pull complete 9982991b8208: Pull complete cb374f741579: Pull complete 732dbfa4c42c: Pull complete 64bbe06a1892: Pull complete 15263a6e39d7: Pull complete d1ebeb495bd3: Pull complete 460984e65c46: Pull complete 109a63e20d99: Pull complete e3c52c6cbb5c: Pull complete Digest: sha256:d5843e74790da22bd3d84d098dcbe21c7318a7c7c4db6bf1de09cb193e64c513 Status: Downloaded newer image for projects.registry.vmware.com/antrea/golang:1.19-nanoserver projects.registry.vmware.com/antrea/golang:1.19-nanoserver ===> Building Antrea bins and antrea/antrea-windows Docker image <=== docker build --pull -t antrea/antrea-windows:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.windows --network host --build-arg GO_VERSION=1.19 --build-arg CNI_BINARIES_VERSION=v1.1.1 --build-arg NANOSERVER_VERSION=1809 --build-arg WIN_BUILD_TAG=caa8cac3d3 . Sending build context to Docker daemon 558.6MB Step 1/21 : ARG WIN_BUILD_TAG Step 2/21 : ARG NANOSERVER_VERSION Step 3/21 : FROM antrea/base-windows:${WIN_BUILD_TAG} as antrea-build-windows caa8cac3d3: Pulling from antrea/base-windows 5ead999142ec: Pulling fs layer 3724c0f91ae7: Pulling fs layer 97a8c2800739: Pulling fs layer c038a2fdb0b5: Pulling fs layer 5ef04caec17a: Pulling fs layer 8bca1755b69d: Pulling fs layer 3ebc2faa6c3e: Pulling fs layer d85edb0b9152: Pulling fs layer b343c39a6010: Pulling fs layer b673c002e661: Pulling fs layer 99b72f4231b5: Pulling fs layer d85edb0b9152: Waiting 3ebc2faa6c3e: Waiting 8bca1755b69d: Waiting c038a2fdb0b5: Waiting b343c39a6010: Waiting 5ef04caec17a: Waiting b673c002e661: Waiting 99b72f4231b5: Waiting 97a8c2800739: Download complete 3724c0f91ae7: Download complete 5ead999142ec: Verifying Checksum 5ead999142ec: Download complete c038a2fdb0b5: Download complete 5ef04caec17a: Verifying Checksum 5ef04caec17a: Download complete 8bca1755b69d: Verifying Checksum 8bca1755b69d: Download complete b343c39a6010: Verifying Checksum b343c39a6010: Download complete 3ebc2faa6c3e: Verifying Checksum 3ebc2faa6c3e: Download complete 99b72f4231b5: Verifying Checksum 99b72f4231b5: Download complete b673c002e661: Verifying Checksum b673c002e661: Download complete d85edb0b9152: Verifying Checksum d85edb0b9152: Download complete 5ead999142ec: Pull complete 3724c0f91ae7: Pull complete 97a8c2800739: Pull complete c038a2fdb0b5: Pull complete 5ef04caec17a: Pull complete 8bca1755b69d: Pull complete 3ebc2faa6c3e: Pull complete d85edb0b9152: Pull complete b343c39a6010: Pull complete b673c002e661: Pull complete 99b72f4231b5: Pull complete Digest: sha256:c6edebf1a80553a5d474b6663657e69bc5b2cf54b116a1759e4d053c609a575d Status: Downloaded newer image for antrea/base-windows:caa8cac3d3 ---> 79d3453071f3 Step 4/21 : WORKDIR /antrea ---> Running in 8019bfcc22c9 Removing intermediate container 8019bfcc22c9 ---> 6f713bbf1027 Step 5/21 : COPY go.mod /antrea/go.mod ---> 30bbece0963c Step 6/21 : RUN go mod download ---> Running in 7a09a9c079f4 Removing intermediate container 7a09a9c079f4 ---> af9b4babd43e Step 7/21 : COPY . /antrea ---> 60d82c848851 Step 8/21 : RUN sh -c 'make windows-bin' ---> Running in aab89dd36ca9 GOOS=windows CGO_ENABLED=0 go build -o C:/antrea/bin -ldflags ' -X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni antrea.io/antrea/cmd/antrea-agent antrea.io/antrea/cmd/antctl Removing intermediate container aab89dd36ca9 ---> a06472a57304 Step 9/21 : FROM mcr.microsoft.com/powershell:lts-nanoserver-${NANOSERVER_VERSION} lts-nanoserver-1809: Pulling from powershell af0153d864f1: Pulling fs layer eb9aadfb9ebf: Pulling fs layer 698ae59174d5: Pulling fs layer ca1ef8c236fc: Pulling fs layer e3789ff6b796: Pulling fs layer a9a89a12f338: Pulling fs layer ccb30ac1f365: Pulling fs layer 0b2e07545fa8: Pulling fs layer 72408742a0d5: Pulling fs layer ca1ef8c236fc: Waiting 0b2e07545fa8: Waiting 72408742a0d5: Waiting e3789ff6b796: Waiting a9a89a12f338: Waiting ccb30ac1f365: Waiting eb9aadfb9ebf: Verifying Checksum eb9aadfb9ebf: Download complete 698ae59174d5: Verifying Checksum 698ae59174d5: Download complete ca1ef8c236fc: Verifying Checksum ca1ef8c236fc: Download complete e3789ff6b796: Verifying Checksum e3789ff6b796: Download complete af0153d864f1: Verifying Checksum af0153d864f1: Download complete ccb30ac1f365: Verifying Checksum ccb30ac1f365: Download complete 0b2e07545fa8: Verifying Checksum 0b2e07545fa8: Download complete 72408742a0d5: Verifying Checksum 72408742a0d5: Download complete a9a89a12f338: Verifying Checksum a9a89a12f338: Download complete af0153d864f1: Pull complete eb9aadfb9ebf: Pull complete 698ae59174d5: Pull complete ca1ef8c236fc: Pull complete e3789ff6b796: Pull complete a9a89a12f338: Pull complete ccb30ac1f365: Pull complete 0b2e07545fa8: Pull complete 72408742a0d5: Pull complete Digest: sha256:41643cdb0d064e3a282badd7c1568e8ffa904536a3e95116ce353696a5d1d471 Status: Downloaded newer image for mcr.microsoft.com/powershell:lts-nanoserver-1809 ---> d57d56a6e450 Step 10/21 : SHELL ["pwsh", "-NoLogo", "-Command", "$ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';"] ---> Running in 494b3abb7438 Removing intermediate container 494b3abb7438 ---> ad61f45fa627 Step 11/21 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>" ---> Running in 3c3062b996af Removing intermediate container 3c3062b996af ---> 4362e1e85c5d Step 12/21 : LABEL description="A Docker image to deploy the Antrea CNI." ---> Running in 950d9303ee7c Removing intermediate container 950d9303ee7c ---> a6d107b46ad6 Step 13/21 : USER ContainerAdministrator ---> Running in 08026f87f27e Removing intermediate container 08026f87f27e ---> 5ed0dd7211b1 Step 14/21 : RUN mkdir -Force C:\k\antrea\bin ---> Running in 53264a32f2da Directory: C:\k\antrea Mode LastWriteTime Length Name ---- ------------- ------ ---- d---- 3/8/2023 8:54 PM bin Removing intermediate container 53264a32f2da ---> 744d452bdcef Step 15/21 : COPY --from=antrea-build-windows /opt/cni/bin /k/antrea/cni ---> da2f7aea09d2 Step 16/21 : COPY --from=antrea-build-windows /antrea/build/images/scripts/Install-WindowsCNI.ps1 /k/antrea/ ---> b3509d1238ca Step 17/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-agent.exe /k/antrea/bin/ ---> c75b95fbe770 Step 18/21 : COPY --from=antrea-build-windows /antrea/bin/antctl.exe /k/antrea/bin/antctl.exe ---> a6560ddef3ad Step 19/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-cni.exe /k/antrea/cni/antrea.exe ---> ddf94081e150 Step 20/21 : RUN mkdir C:\k\antrea\utils ---> Running in 5e156644b55e Directory: C:\k\antrea Mode LastWriteTime Length Name ---- ------------- ------ ---- d---- 3/8/2023 8:55 PM utils Removing intermediate container 5e156644b55e ---> db43d08befea Step 21/21 : COPY --from=antrea-build-windows /wins/wins.exe /k/antrea/utils/wins.exe ---> f84ef4d5fd9d [Warning] One or more build-args [CNI_BINARIES_VERSION GO_VERSION] were not consumed Successfully built f84ef4d5fd9d Successfully tagged antrea/antrea-windows:v1.11.0-dev-75a53913.dirty docker tag antrea/antrea-windows:v1.11.0-dev-75a53913.dirty antrea/antrea-windows ==== Reverting Windows VM a-ms-0008-win-1 ===== Windows VM a-ms-0008-win-1 powered on The command completed successfully. projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.13 projects.registry.vmware.com/antrea/agnhost:2.29 projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0 projects.registry.vmware.com/antrea/e2eteam-pause:3.2 mcr.microsoft.com/windows/servercore/iis:latest The image projects.registry.vmware.com/antrea/antrea-windows:latest already exists, renaming the old one with ID sha256:4653103584af17d23287e7ce178d037e588f8188506402323dc0dff943264b2b to empty string Loaded image: projects.registry.vmware.com/antrea/antrea-windows:latest ====== Running Antrea e2e Tests ====== customresourcedefinition.apiextensions.k8s.io/antreaagentinfos.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/antreacontrollerinfos.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/clustergroups.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/clusternetworkpolicies.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/egresses.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/externalentities.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/externalippools.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/externalnodes.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/ippools.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/supportbundlecollections.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/tiers.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/traceflows.crd.antrea.io created customresourcedefinition.apiextensions.k8s.io/trafficcontrols.crd.antrea.io created serviceaccount/antrea-agent created serviceaccount/antctl created serviceaccount/antrea-controller created secret/antrea-agent-service-account-token created secret/antctl-service-account-token created configmap/antrea-config created customresourcedefinition.apiextensions.k8s.io/groups.crd.antrea.io created clusterrole.rbac.authorization.k8s.io/antrea-agent created clusterrole.rbac.authorization.k8s.io/antctl created clusterrole.rbac.authorization.k8s.io/antrea-cluster-identity-reader created clusterrole.rbac.authorization.k8s.io/antrea-controller created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-edit created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-view created clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-edit created clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-view created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-edit created clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-view created clusterrolebinding.rbac.authorization.k8s.io/antrea-agent created clusterrolebinding.rbac.authorization.k8s.io/antctl created clusterrolebinding.rbac.authorization.k8s.io/antrea-controller created service/antrea created daemonset.apps/antrea-agent created deployment.apps/antrea-controller created apiservice.apiregistration.k8s.io/v1beta2.controlplane.antrea.io created apiservice.apiregistration.k8s.io/v1beta1.system.antrea.io created apiservice.apiregistration.k8s.io/v1alpha1.stats.antrea.io created mutatingwebhookconfiguration.admissionregistration.k8s.io/crdmutator.antrea.io created validatingwebhookconfiguration.admissionregistration.k8s.io/crdvalidator.antrea.io created configmap/antrea-agent-windows-kht6m7hthm created configmap/antrea-windows-config-db2774h9dt created daemonset.apps/antrea-agent-windows created deployment.apps/coredns restarted Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated... Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated... Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated... Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination... Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination... Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination... Waiting for deployment "coredns" rollout to finish: 1 of 2 updated replicas are available... deployment "coredns" successfully rolled out deployment "antrea-controller" successfully rolled out daemon set "antrea-agent" successfully rolled out Waiting for daemon set "antrea-agent-windows" rollout to finish: 0 of 2 updated pods are available... Waiting for daemon set "antrea-agent-windows" rollout to finish: 1 of 2 updated pods are available... daemon set "antrea-agent-windows" successfully rolled out Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-BF-B1 10 Gbps Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-B7-EE 10 Gbps === Generate ssh-config === Generating ssh-config for Node a-ms-0008-0 Generating ssh-config for Node a-ms-0008-1 Generating ssh-config for Node a-ms-0008-win-0 Generating ssh-config for Node a-ms-0008-win-1 ====== Run test with e2e test ====== ... ```
test
windows test used stale image on one node describe the bug a clear and concise description of what the bug is if you believe this bug is a security issue please don t use this template and follow our security md when i was debugging an issue on a windows testbed i found one antrea agent windows pod ran an old image the correct version should be dev dirty antrea agent windows nrpvd agent go starting antrea agent version dev dirty antrea agent windows ztlsr agent go starting antrea agent version dev dirty testbed deployment log bin sh xe tmp sh set e head ci docker registry docker registry projects registry vmware com sed i t a skipifhaswindowsnodes t test nodeportlocal test go ci jenkins test sh testcase windows registry projects registry vmware com proxyall clean up stale files folders older than days under tmp cleanup antrea installation no resources found in antrea test namespace building antrea for the following commit commit merge author kumar atish date wed mar merge into pkg agent apiserver handlers memberlist handler go pkg agent apiserver handlers memberlist handler test go pkg support dump go deleted images untagged antrea cni binaries untagged projects registry vmware com antrea cni binaries untagged projects registry vmware com antrea cni binaries deleted deleted deleted deleted deleted deleted deleted untagged ubuntu untagged projects registry vmware com antrea ubuntu untagged projects registry vmware com antrea ubuntu untagged agnhost untagged projects registry vmware com antrea agnhost untagged projects registry vmware com antrea agnhost deleted deleted deleted deleted deleted deleted deleted deleted deleted deleted untagged antrea openvswitch debs antrea untagged projects registry vmware com antrea openvswitch debs antrea untagged projects registry vmware com antrea openvswitch debs deleted deleted deleted deleted deleted deleted deleted untagged nginx alpine untagged projects registry vmware com antrea nginx alpine untagged projects registry vmware com antrea nginx deleted deleted deleted deleted deleted untagged antrea openvswitch antrea untagged projects registry vmware com antrea openvswitch antrea untagged projects registry vmware com antrea openvswitch untagged antrea base ubuntu antrea untagged projects registry vmware com antrea base ubuntu antrea untagged projects registry vmware com antrea base ubuntu total reclaimed space build tag antrea pulling from antrea ubuntu digest status downloaded newer image for projects registry vmware com antrea ubuntu projects registry vmware com antrea ubuntu pulling from antrea golang digest status image is up to date for projects registry vmware com antrea golang projects registry vmware com antrea golang antrea pulling from antrea openvswitch debs already exists pulling fs layer pulling fs layer pulling fs layer verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete digest status downloaded newer image for projects registry vmware com antrea openvswitch debs antrea projects registry vmware com antrea openvswitch debs antrea antrea pulling from antrea openvswitch digest status downloaded newer image for projects registry vmware com antrea openvswitch antrea projects registry vmware com antrea openvswitch antrea pulling from antrea cni binaries already exists pulling fs layer pulling fs layer verifying checksum download complete verifying checksum download complete pull complete pull complete digest status downloaded newer image for projects registry vmware com antrea cni binaries projects registry vmware com antrea cni binaries antrea pulling from antrea base ubuntu digest status downloaded newer image for projects registry vmware com antrea base ubuntu antrea projects registry vmware com antrea base ubuntu antrea sending build context to docker daemon step from ubuntu as ovs debs step arg ovs version using cache step run apt get update debian frontend noninteractive apt get install y no install recommends wget curl git ca certificates build essential fakeroot graphviz autoconf automake debhelper dh python dh autoreconf libssl dev libtool openssl procps all twisted zope interface sphinx libunbound dev using cache step copy apply patches sh using cache step run wget q o tar xz c tmp cd tmp openvswitch apply patches sh deb build options parallel nocheck fakeroot debian rules binary cd tmp mkdir ovs debs mv libopenvswitch deb openvswitch common deb openvswitch switch deb python openvswitch deb openvswitch ipsec deb ovs debs cd rm rf tmp openvswitch using cache successfully built successfully tagged antrea openvswitch debs antrea sending build context to docker daemon step from ubuntu as ovs debs step arg ovs version using cache step run apt get update debian frontend noninteractive apt get install y no install recommends wget curl git ca certificates build essential fakeroot graphviz autoconf automake debhelper dh python dh autoreconf libssl dev libtool openssl procps all twisted zope interface sphinx libunbound dev using cache step copy apply patches sh using cache step run wget q o tar xz c tmp cd tmp openvswitch apply patches sh deb build options parallel nocheck fakeroot debian rules binary cd tmp mkdir ovs debs mv libopenvswitch deb openvswitch common deb openvswitch switch deb python openvswitch deb openvswitch ipsec deb ovs debs cd rm rf tmp openvswitch using cache step from ubuntu step label maintainer antrea using cache step label description a docker image based on ubuntu which includes open vswitch built from source using cache step copy from ovs debs tmp ovs debs tmp ovs debs using cache step copy charon logging conf tmp using cache step run apt get update apt get install y no install recommends iptables logrotate libstrongswan standard plugins dpkg i tmp ovs debs deb apt get f y no install recommends install rm rf var cache apt var lib apt lists sed i rotate a size etc logrotate d openvswitch switch sed i filelog r tmp charon logging conf etc strongswan d charon logging conf rm rf tmp using cache successfully built successfully tagged antrea openvswitch antrea var lib jenkins workspace antrea windows proxyall for pull request sending build context to docker daemon step arg build tag step from ubuntu as cni binaries step arg cni binaries version using cache step arg whereabouts version using cache step run apt get update apt get install y no install recommends wget ca certificates using cache step env cni plugins host local loopback portmap bandwidth using cache step run set eux dpkgarch dpkg print architecture case dpkgarch in pluginsarch armhf pluginsarch arm pluginsarch pluginsarch echo echo unsupported architecture dpkgarch echo exit esac mkdir p opt cni bin wget q o tar xz c opt cni bin cni plugins wget q o opt cni bin whereabouts chmod x opt cni bin whereabouts using cache successfully built successfully tagged antrea cni binaries sending build context to docker daemon step arg build tag step from ubuntu as cni binaries step arg cni binaries version using cache step arg whereabouts version using cache step run apt get update apt get install y no install recommends wget ca certificates using cache step env cni plugins host local loopback portmap bandwidth using cache step run set eux dpkgarch dpkg print architecture case dpkgarch in pluginsarch armhf pluginsarch arm pluginsarch pluginsarch echo echo unsupported architecture dpkgarch echo exit esac mkdir p opt cni bin wget q o tar xz c opt cni bin cni plugins wget q o opt cni bin whereabouts chmod x opt cni bin whereabouts using cache step from antrea openvswitch build tag step arg suricata version using cache step label maintainer antrea using cache step label description an ubuntu based docker base image for antrea using cache step user root using cache step add iptables wrapper installer sh using cache step run apt get update apt get install y no install recommends ipset jq inotify tools gpg agent software properties common add apt repository ppa oisf suricata suricata version apt get update apt get install y suricata apt get remove y gpg agent software properties common apt get autoremove y rm rf var cache apt var lib apt lists chmod x iptables wrapper installer sh iptables wrapper installer sh using cache step copy from cni binaries opt cni bin opt cni bin using cache successfully built successfully tagged antrea base ubuntu antrea var lib jenkins workspace antrea windows proxyall for pull request building antrea bins and antrea antrea ubuntu docker image docker build t antrea antrea ubuntu dev dirty f build images dockerfile build ubuntu build arg ovs version build arg go version build arg build tag antrea sending build context to docker daemon step arg go version step arg build tag step from golang go version as antrea build step workdir antrea running in removing intermediate container step copy go mod antrea go mod step run go mod download running in removing intermediate container step copy antrea step run make antrea agent antrea controller antrea cni antctl linux running in goos linux go build o antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea agent goos linux go build o antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea controller goos linux cgo enabled go build o antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea cni removing intermediate container step run mv bin antctl linux bin antctl running in removing intermediate container step from antrea base ubuntu build tag step label maintainer antrea using cache step label description the docker image to deploy the antrea cni using cache step user root using cache step copy build images scripts usr local bin using cache step copy from antrea build antrea bin usr local bin one or more build args were not consumed successfully built successfully tagged antrea antrea ubuntu dev dirty docker tag antrea antrea ubuntu dev dirty antrea antrea ubuntu delivering antrea to all nodes updating yaml files to enable proxyall pull necessary images on control plane node projects registry vmware com antrea agnhost projects registry vmware com antrea nginx alpine deliver antrea to linux worker nodes and pull necessary images on worker nodes sending incremental file list antrea ubuntu tar s s s s s s s s xfr to chk sent bytes received bytes bytes sec total size is speedup is deleted images untagged gcr io test images agnhost untagged gcr io test images agnhost deleted deleted deleted deleted deleted deleted deleted deleted deleted deleted untagged agnhost untagged projects registry vmware com antrea agnhost untagged projects registry vmware com antrea agnhost deleted deleted deleted deleted deleted deleted deleted deleted deleted deleted untagged nginx alpine untagged projects registry vmware com antrea nginx alpine untagged projects registry vmware com antrea nginx deleted deleted deleted deleted deleted total reclaimed space the image antrea antrea ubuntu latest already exists renaming the old one with id to empty string loaded image antrea antrea ubuntu latest projects registry vmware com antrea agnhost projects registry vmware com antrea nginx alpine gcr io test images agnhost deliver antrea windows to windows worker nodes and pull necessary images on windows worker nodes reverting windows vm a ms win windows vm a ms win powered on projects registry vmware com antrea sigwindowstools kube proxy projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea jessie dnsutils projects registry vmware com antrea pause error response from daemon get certificate has expired or is not yet valid build windows on windows node nanoserver pulling from antrea golang pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer waiting waiting waiting waiting waiting waiting waiting verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for projects registry vmware com antrea golang nanoserver projects registry vmware com antrea golang nanoserver building antrea bins and antrea antrea windows docker image docker build pull t antrea antrea windows dev dirty f build images dockerfile build windows network host build arg go version build arg cni binaries version build arg nanoserver version build arg win build tag sending build context to docker daemon step arg win build tag step arg nanoserver version step from antrea base windows win build tag as antrea build windows pulling from antrea base windows pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer waiting waiting waiting waiting waiting waiting waiting waiting download complete download complete verifying checksum download complete download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for antrea base windows step workdir antrea running in removing intermediate container step copy go mod antrea go mod step run go mod download running in removing intermediate container step copy antrea step run sh c make windows bin running in goos windows cgo enabled go build o c antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea cni antrea io antrea cmd antrea agent antrea io antrea cmd antctl removing intermediate container step from mcr microsoft com powershell lts nanoserver nanoserver version lts nanoserver pulling from powershell pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer waiting waiting waiting waiting waiting waiting verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for mcr microsoft com powershell lts nanoserver step shell running in removing intermediate container step label maintainer antrea running in removing intermediate container step label description a docker image to deploy the antrea cni running in removing intermediate container step user containeradministrator running in removing intermediate container step run mkdir force c k antrea bin running in directory c k antrea  lastwritetime length name   d pm bin removing intermediate container step copy from antrea build windows opt cni bin k antrea cni step copy from antrea build windows antrea build images scripts install windowscni k antrea step copy from antrea build windows antrea bin antrea agent exe k antrea bin step copy from antrea build windows antrea bin antctl exe k antrea bin antctl exe step copy from antrea build windows antrea bin antrea cni exe k antrea cni antrea exe step run mkdir c k antrea utils running in directory c k antrea  lastwritetime length name   d pm utils removing intermediate container step copy from antrea build windows wins wins exe k antrea utils wins exe one or more build args were not consumed successfully built successfully tagged antrea antrea windows dev dirty docker tag antrea antrea windows dev dirty antrea antrea windows reverting windows vm a ms win windows vm a ms win powered on the command completed successfully projects registry vmware com antrea sigwindowstools kube proxy projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea jessie dnsutils projects registry vmware com antrea pause mcr microsoft com windows servercore iis latest the image projects registry vmware com antrea antrea windows latest already exists renaming the old one with id to empty string loaded image projects registry vmware com antrea antrea windows latest running antrea tests customresourcedefinition apiextensions io antreaagentinfos crd antrea io created customresourcedefinition apiextensions io antreacontrollerinfos crd antrea io created customresourcedefinition apiextensions io clustergroups crd antrea io created customresourcedefinition apiextensions io clusternetworkpolicies crd antrea io created customresourcedefinition apiextensions io egresses crd antrea io created customresourcedefinition apiextensions io externalentities crd antrea io created customresourcedefinition apiextensions io externalippools crd antrea io created customresourcedefinition apiextensions io externalnodes crd antrea io created customresourcedefinition apiextensions io ippools crd antrea io created customresourcedefinition apiextensions io networkpolicies crd antrea io created customresourcedefinition apiextensions io supportbundlecollections crd antrea io created customresourcedefinition apiextensions io tiers crd antrea io created customresourcedefinition apiextensions io traceflows crd antrea io created customresourcedefinition apiextensions io trafficcontrols crd antrea io created serviceaccount antrea agent created serviceaccount antctl created serviceaccount antrea controller created secret antrea agent service account token created secret antctl service account token created configmap antrea config created customresourcedefinition apiextensions io groups crd antrea io created clusterrole rbac authorization io antrea agent created clusterrole rbac authorization io antctl created clusterrole rbac authorization io antrea cluster identity reader created clusterrole rbac authorization io antrea controller created clusterrole rbac authorization io aggregate antrea policies edit created clusterrole rbac authorization io aggregate antrea policies view created clusterrole rbac authorization io aggregate traceflows edit created clusterrole rbac authorization io aggregate traceflows view created clusterrole rbac authorization io aggregate antrea clustergroups edit created clusterrole rbac authorization io aggregate antrea clustergroups view created clusterrolebinding rbac authorization io antrea agent created clusterrolebinding rbac authorization io antctl created clusterrolebinding rbac authorization io antrea controller created service antrea created daemonset apps antrea agent created deployment apps antrea controller created apiservice apiregistration io controlplane antrea io created apiservice apiregistration io system antrea io created apiservice apiregistration io stats antrea io created mutatingwebhookconfiguration admissionregistration io crdmutator antrea io created validatingwebhookconfiguration admissionregistration io crdvalidator antrea io created configmap antrea agent windows created configmap antrea windows config created daemonset apps antrea agent windows created deployment apps coredns restarted waiting for deployment coredns rollout to finish out of new replicas have been updated waiting for deployment coredns rollout to finish out of new replicas have been updated waiting for deployment coredns rollout to finish out of new replicas have been updated waiting for deployment coredns rollout to finish old replicas are pending termination waiting for deployment coredns rollout to finish old replicas are pending termination waiting for deployment coredns rollout to finish old replicas are pending termination waiting for deployment coredns rollout to finish of updated replicas are available deployment coredns successfully rolled out deployment antrea controller successfully rolled out daemon set antrea agent successfully rolled out waiting for daemon set antrea agent windows rollout to finish of updated pods are available waiting for daemon set antrea agent windows rollout to finish of updated pods are available daemon set antrea agent windows successfully rolled out name interfacedescription ifindex status macaddress linkspeed br int hyper v virtual ethernet adapter up bf gbps name interfacedescription ifindex status macaddress linkspeed br int hyper v virtual ethernet adapter up ee gbps generate ssh config generating ssh config for node a ms generating ssh config for node a ms generating ssh config for node a ms win generating ssh config for node a ms win run test with test
1
80,598
7,751,829,219
IssuesEvent
2018-05-30 18:19:28
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
Test failure: GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd
test-run-uwp-coreclr
Opened on behalf of @Sunny-pu The test `GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd` has failed. Test Infrastructure Failure: Could not load file or assembly 'System.Memory, Version=4.1.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified. Expected: True Actual: False Stack Trace: Build : Master - 20180528.01 (Ready-To-Run Tests) Failing configurations: - ubuntu.1604.amd64 - x64 Details: https://mc.dot.net/#/product/netcore/master/source/official~2Fcoreclr~2Fmaster~2F/type/test~2Ffunctional~2Fr2r~2Fcli~2F/build/20180528.01/workItem/GC.Coverage.XUnitWrapper/analysis/xunit/GC_Coverage._delete_next_card_table_delete_next_card_table_~2F_delete_next_card_table_delete_next_card_table_cmd
1.0
Test failure: GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd - Opened on behalf of @Sunny-pu The test `GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd` has failed. Test Infrastructure Failure: Could not load file or assembly 'System.Memory, Version=4.1.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified. Expected: True Actual: False Stack Trace: Build : Master - 20180528.01 (Ready-To-Run Tests) Failing configurations: - ubuntu.1604.amd64 - x64 Details: https://mc.dot.net/#/product/netcore/master/source/official~2Fcoreclr~2Fmaster~2F/type/test~2Ffunctional~2Fr2r~2Fcli~2F/build/20180528.01/workItem/GC.Coverage.XUnitWrapper/analysis/xunit/GC_Coverage._delete_next_card_table_delete_next_card_table_~2F_delete_next_card_table_delete_next_card_table_cmd
test
test failure gc coverage delete next card table delete next card table delete next card table delete next card table cmd opened on behalf of sunny pu the test gc coverage delete next card table delete next card table delete next card table delete next card table cmd has failed test infrastructure failure could not load file or assembly system memory version culture neutral publickeytoken the system cannot find the file specified expected true actual false stack trace build master ready to run tests failing configurations ubuntu details
1
132,796
10,764,165,103
IssuesEvent
2019-11-01 07:18:35
a2000-erp-team/WEBERP
https://api.github.com/repos/a2000-erp-team/WEBERP
opened
Procurement-Admin-Item Account Group-Edit. Mapping for "Stock A/c" shows Stock Adjust A/c. Kindly correct because misleading, cannot find where to change the Stock A/c. mapping.
WEB ERP Testing By Katrina
![image.png](https://images.zenhubusercontent.com/5cf8a04f2e4fe4691d7f073e/fe92d5ec-50fd-4b72-a8e3-a2b0499ba9d9)
1.0
Procurement-Admin-Item Account Group-Edit. Mapping for "Stock A/c" shows Stock Adjust A/c. Kindly correct because misleading, cannot find where to change the Stock A/c. mapping. - ![image.png](https://images.zenhubusercontent.com/5cf8a04f2e4fe4691d7f073e/fe92d5ec-50fd-4b72-a8e3-a2b0499ba9d9)
test
procurement admin item account group edit mapping for stock a c shows stock adjust a c kindly correct because misleading cannot find where to change the stock a c mapping
1
302,862
26,169,264,911
IssuesEvent
2023-01-01 17:58:02
LAION-AI/Open-Assistant
https://api.github.com/repos/LAION-AI/Open-Assistant
closed
Contract tests between API and Discord Bot
discord-bot testing
As per discussion in #180, the communication between the discord bot and API should be tested through a Contract test. This test should test that each call satisfies the OpenAPI schema inputs and endpoints. That way, changes to the endpoint contract will automatically be caught by these tests.
1.0
Contract tests between API and Discord Bot - As per discussion in #180, the communication between the discord bot and API should be tested through a Contract test. This test should test that each call satisfies the OpenAPI schema inputs and endpoints. That way, changes to the endpoint contract will automatically be caught by these tests.
test
contract tests between api and discord bot as per discussion in the communication between the discord bot and api should be tested through a contract test this test should test that each call satisfies the openapi schema inputs and endpoints that way changes to the endpoint contract will automatically be caught by these tests
1
63,849
6,886,277,634
IssuesEvent
2017-11-21 18:53:56
RetroWoW/PTR
https://api.github.com/repos/RetroWoW/PTR
closed
T2 Paladin 8/8 Bonus not scaling with spelldmg any more
Class - Paladin Requires Testing / Confirmation
**Description**: With the new core the 8/8 set bonus does exactly the dmg that it says on the tooltip and doesn't scale with spelldmg any more **Expected behavior**: Should scale
1.0
T2 Paladin 8/8 Bonus not scaling with spelldmg any more - **Description**: With the new core the 8/8 set bonus does exactly the dmg that it says on the tooltip and doesn't scale with spelldmg any more **Expected behavior**: Should scale
test
paladin bonus not scaling with spelldmg any more description with the new core the set bonus does exactly the dmg that it says on the tooltip and doesn t scale with spelldmg any more expected behavior should scale
1
317,155
27,217,199,084
IssuesEvent
2023-02-20 23:35:57
acikkaynak/deprem-yardim-frontend
https://api.github.com/repos/acikkaynak/deprem-yardim-frontend
closed
bug: Can't click scan field button when map layer window is open
bug discussion approved ios android p0 test-failed
## Bug Definition Can't click scan field button when map layer window is open -- Please add discord contact information ** discord username: @afetharita#0001 ** ## Bug environment Describe the environment produces the bug. For example: rc.afetharita.com ## Describe how you are producing the bug step by step 1. Go to '..RC.'rc.afetharita.com 2. Click map layer button 3. Click to scan field button 4. Bug appears ## Expected Behaviour map layers should be closed and scan layer shold be clickable ## Screen shots If possible, add screenshots to describe your bug. ## Desktop Information - Operating System: Windows 10 - Browser chrome - Version 110.0.5481.97 (Resmi Derleme) (64 bit) Sürümü ## Mobile Phone Information - Devıce: [for example iPhone6] - Operating System (with the version): [for example iOS8.1] - Version [for example default browser, safari] - Browser Version [for example 22] ## Additional Context Add any other context about the bug here
1.0
bug: Can't click scan field button when map layer window is open - ## Bug Definition Can't click scan field button when map layer window is open -- Please add discord contact information ** discord username: @afetharita#0001 ** ## Bug environment Describe the environment produces the bug. For example: rc.afetharita.com ## Describe how you are producing the bug step by step 1. Go to '..RC.'rc.afetharita.com 2. Click map layer button 3. Click to scan field button 4. Bug appears ## Expected Behaviour map layers should be closed and scan layer shold be clickable ## Screen shots If possible, add screenshots to describe your bug. ## Desktop Information - Operating System: Windows 10 - Browser chrome - Version 110.0.5481.97 (Resmi Derleme) (64 bit) Sürümü ## Mobile Phone Information - Devıce: [for example iPhone6] - Operating System (with the version): [for example iOS8.1] - Version [for example default browser, safari] - Browser Version [for example 22] ## Additional Context Add any other context about the bug here
test
bug can t click scan field button when map layer window is open bug definition can t click scan field button when map layer window is open please add discord contact information discord username afetharita  bug environment describe the environment produces the bug for example rc afetharita com describe how you are producing the bug step by step go to rc rc afetharita com click map layer button click to scan field button bug appears expected behaviour map layers should be closed and scan layer shold be clickable screen shots if possible add screenshots to describe your bug desktop information operating system windows browser chrome version resmi derleme bit sürümü mobile phone information devıce operating system with the version version browser version additional context add any other context about the bug here
1
459,863
13,200,379,420
IssuesEvent
2020-08-14 08:07:10
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
play.google.com - site is not usable
browser-android-components engine-gecko priority-critical
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/56599 --> <!-- @extra_labels: browser-android-components --> **URL**: https://play.google.com/store/apps/details?id%3Dcom.google.android.apps.maps%26referrer%3Dutm_source%253Dhelp-center%2526utm_medium%253Dhc-answer%2526utm_campaign%253Dhelp-center-mg-1250066 **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android **Tested Another Browser**: Yes Safari **Problem type**: Site is not usable **Description**: Unable to login **Steps to Reproduce**: I am hacked <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200808213325</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
play.google.com - site is not usable - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/56599 --> <!-- @extra_labels: browser-android-components --> **URL**: https://play.google.com/store/apps/details?id%3Dcom.google.android.apps.maps%26referrer%3Dutm_source%253Dhelp-center%2526utm_medium%253Dhc-answer%2526utm_campaign%253Dhelp-center-mg-1250066 **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android **Tested Another Browser**: Yes Safari **Problem type**: Site is not usable **Description**: Unable to login **Steps to Reproduce**: I am hacked <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200808213325</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
play google com site is not usable url browser version firefox mobile operating system android tested another browser yes safari problem type site is not usable description unable to login steps to reproduce i am hacked browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
0
113,705
17,150,880,143
IssuesEvent
2021-07-13 20:25:42
snowdensb/braindump
https://api.github.com/repos/snowdensb/braindump
opened
CVE-2019-18797 (Medium) detected in node-sass-3.12.1.tgz, libsass3.3.6
security vulnerability
## CVE-2019-18797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-3.12.1.tgz</b>, <b>libsass3.3.6</b></p></summary> <p> <details><summary><b>node-sass-3.12.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz</a></p> <p>Path to dependency file: braindump/package.json</p> <p>Path to vulnerable library: braindump/node_modules/node-sass</p> <p> Dependency Hierarchy: - gulp-sass-2.3.2.tgz (Root Library) - :x: **node-sass-3.12.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/braindump/commit/815ae0afebcf867f02143f3ab9cf88b1d4dacdec">815ae0afebcf867f02143f3ab9cf88b1d4dacdec</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp. <p>Publish Date: 2019-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p> <p>Release Date: 2019-11-06</p> <p>Fix Resolution: LibSass - 3.6.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"3.12.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp-sass:2.3.2;node-sass:3.12.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"LibSass - 3.6.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-18797","vulnerabilityDetails":"LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-18797 (Medium) detected in node-sass-3.12.1.tgz, libsass3.3.6 - ## CVE-2019-18797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-3.12.1.tgz</b>, <b>libsass3.3.6</b></p></summary> <p> <details><summary><b>node-sass-3.12.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz</a></p> <p>Path to dependency file: braindump/package.json</p> <p>Path to vulnerable library: braindump/node_modules/node-sass</p> <p> Dependency Hierarchy: - gulp-sass-2.3.2.tgz (Root Library) - :x: **node-sass-3.12.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/braindump/commit/815ae0afebcf867f02143f3ab9cf88b1d4dacdec">815ae0afebcf867f02143f3ab9cf88b1d4dacdec</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp. <p>Publish Date: 2019-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p> <p>Release Date: 2019-11-06</p> <p>Fix Resolution: LibSass - 3.6.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"3.12.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp-sass:2.3.2;node-sass:3.12.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"LibSass - 3.6.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-18797","vulnerabilityDetails":"LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in node sass tgz cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file braindump package json path to vulnerable library braindump node modules node sass dependency hierarchy gulp sass tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree gulp sass node sass isminimumfixversionavailable true minimumfixversion libsass basebranches vulnerabilityidentifier cve vulnerabilitydetails libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp vulnerabilityurl
0
772,401
27,120,677,639
IssuesEvent
2023-02-15 22:33:56
Lightning-AI/lightning
https://api.github.com/repos/Lightning-AI/lightning
closed
QAT callback doesn't work with multi-input models or forward kwargs
bug help wanted 3rd party priority: 1
## 🐛 Bug The `QuantizationAwareTraining` can only handle models whose forward signature looks like `forward(data)`. Adding additional positional or keyword arguments breaks the forward propagation. ## Please reproduce using the BoringModel The `BoringModel` doesn't have the right forward signature to produce errors, but I included an equivalently simple model below ### To Reproduce ```python import functools import torch import torch.nn as nn class Simple(nn.Module): def __init__(self): super(Simple, self).__init__() self.conv = nn.Conv2d(3, 16, kernel_size=3) self.softmax = nn.Softmax(dim=3) def forward(self, a, b, softmax=False): """Use 2 positional arguments and 1 keyword argument""" feat_b = self.conv(b) feat_a = self.conv(a) combined = feat_a + feat_b if softmax: return self.softmax(combined) else: return combined @property def example_input_array(self): return torch.randn(1, 3, 28, 28), torch.randn(1, 3, 28, 28) def wrap_quantize_forward_context(model, func): """From QuantizationAwareTraining callback""" @functools.wraps(func) def wrapper(data): data = model.quant(data) data = func(data) data = model.dequant(data) return data return wrapper # Works fine without quantization model = Simple() input_sample = model.example_input_array _ = model(*input_sample) _ = model(*input_sample, softmax=True) # Simulate the `on_fit_start` hook behavior model.quant = torch.quantization.QuantStub() model.dequant = torch.quantization.DeQuantStub() model.forward = wrap_quantize_forward_context(model, model.forward) _ = model(*input_sample) _ = model(*input_sample, softmax=True) ``` The last two lines will raise the following exceptions ``` TypeError: wrapper() takes 1 positional argument but 2 were given TypeError: wrapper() got an unexpected keyword argument 'softmax' ``` ### Expected behavior Forward prop should work fine for both the quantized and unquantized model ### Environment * CUDA: - GPU: GeForce GTX 1060 with Max-Q Design - available: True - version: 10.2 * Packages: - numpy: 1.19.2 - pyTorch_debug: False - pyTorch_version: 1.8.1+cu102 - pytorch-lightning: 1.3.0 - tqdm: 4.59.0 * System: - OS: Linux - architecture: 64bit - processor: x86_64 - python: 3.6.13 - version: #148-Ubuntu SMP Sat May 8 02:33:43 UTC 2021 ### Additional context The solution to the minimal example is to make the decoration wrapper more generic ```python def wrap_quantize_forward_context(model, func): """From QuantizationAwareTraining callback""" @functools.wraps(func) def wrapper(*data, **kwargs): data = model.quant(data) data = func(*data, **kwargs) data = model.dequant(data) return data return wrapper ``` However, I notice that `callbacks/quantization.py` also includes a second decorator `wrap_qat_forward_context`, and I am not exactly clear on the intent behind having two separate decorators. I see that `wrap_quantize_forward_context` is only used in the `on_fit_end` hook. Any insight here @Borda? I can make a PR for the above solution, but first I tested it in an actual training by patching my local install ```python def wrap_qat_forward_context( quant_cb, model: 'pl.LightningModule', func: Callable, trigger_condition: Optional[Union[Callable, int]] = None ) -> Callable: @functools.wraps(func) def wrapper(*data, **kwargs) -> Any: # <-- Changes to decoration signature _is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer) _is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition _quant_run = trigger_condition is None or _is_func_true or _is_count_true if _quant_run: quant_cb._forward_calls += 1 data = model.quant(data) # <-- Triggers AttributeError 'tuple' object has no attribute 'detach' data = func(*data, **kwargs) if _quant_run: data = model.dequant(data) return data return wrapper ``` So I tried also expanding the `data` before passing to `model.quant` but this goes back to a `TypeError` ```python def wrap_qat_forward_context( quant_cb, model: 'pl.LightningModule', func: Callable, trigger_condition: Optional[Union[Callable, int]] = None ) -> Callable: @functools.wraps(func) def wrapper(*data, **kwargs) -> Any: _is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer) _is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition _quant_run = trigger_condition is None or _is_func_true or _is_count_true if _quant_run: quant_cb._forward_calls += 1 data = model.quant(*data) # <-- Attempted tuple unpacking, get TypeError: forward() takes 2 positional arguments but 3 were given data = func(*data, **kwargs) if _quant_run: data = model.dequant(*data) return data return wrapper ```
1.0
QAT callback doesn't work with multi-input models or forward kwargs - ## 🐛 Bug The `QuantizationAwareTraining` can only handle models whose forward signature looks like `forward(data)`. Adding additional positional or keyword arguments breaks the forward propagation. ## Please reproduce using the BoringModel The `BoringModel` doesn't have the right forward signature to produce errors, but I included an equivalently simple model below ### To Reproduce ```python import functools import torch import torch.nn as nn class Simple(nn.Module): def __init__(self): super(Simple, self).__init__() self.conv = nn.Conv2d(3, 16, kernel_size=3) self.softmax = nn.Softmax(dim=3) def forward(self, a, b, softmax=False): """Use 2 positional arguments and 1 keyword argument""" feat_b = self.conv(b) feat_a = self.conv(a) combined = feat_a + feat_b if softmax: return self.softmax(combined) else: return combined @property def example_input_array(self): return torch.randn(1, 3, 28, 28), torch.randn(1, 3, 28, 28) def wrap_quantize_forward_context(model, func): """From QuantizationAwareTraining callback""" @functools.wraps(func) def wrapper(data): data = model.quant(data) data = func(data) data = model.dequant(data) return data return wrapper # Works fine without quantization model = Simple() input_sample = model.example_input_array _ = model(*input_sample) _ = model(*input_sample, softmax=True) # Simulate the `on_fit_start` hook behavior model.quant = torch.quantization.QuantStub() model.dequant = torch.quantization.DeQuantStub() model.forward = wrap_quantize_forward_context(model, model.forward) _ = model(*input_sample) _ = model(*input_sample, softmax=True) ``` The last two lines will raise the following exceptions ``` TypeError: wrapper() takes 1 positional argument but 2 were given TypeError: wrapper() got an unexpected keyword argument 'softmax' ``` ### Expected behavior Forward prop should work fine for both the quantized and unquantized model ### Environment * CUDA: - GPU: GeForce GTX 1060 with Max-Q Design - available: True - version: 10.2 * Packages: - numpy: 1.19.2 - pyTorch_debug: False - pyTorch_version: 1.8.1+cu102 - pytorch-lightning: 1.3.0 - tqdm: 4.59.0 * System: - OS: Linux - architecture: 64bit - processor: x86_64 - python: 3.6.13 - version: #148-Ubuntu SMP Sat May 8 02:33:43 UTC 2021 ### Additional context The solution to the minimal example is to make the decoration wrapper more generic ```python def wrap_quantize_forward_context(model, func): """From QuantizationAwareTraining callback""" @functools.wraps(func) def wrapper(*data, **kwargs): data = model.quant(data) data = func(*data, **kwargs) data = model.dequant(data) return data return wrapper ``` However, I notice that `callbacks/quantization.py` also includes a second decorator `wrap_qat_forward_context`, and I am not exactly clear on the intent behind having two separate decorators. I see that `wrap_quantize_forward_context` is only used in the `on_fit_end` hook. Any insight here @Borda? I can make a PR for the above solution, but first I tested it in an actual training by patching my local install ```python def wrap_qat_forward_context( quant_cb, model: 'pl.LightningModule', func: Callable, trigger_condition: Optional[Union[Callable, int]] = None ) -> Callable: @functools.wraps(func) def wrapper(*data, **kwargs) -> Any: # <-- Changes to decoration signature _is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer) _is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition _quant_run = trigger_condition is None or _is_func_true or _is_count_true if _quant_run: quant_cb._forward_calls += 1 data = model.quant(data) # <-- Triggers AttributeError 'tuple' object has no attribute 'detach' data = func(*data, **kwargs) if _quant_run: data = model.dequant(data) return data return wrapper ``` So I tried also expanding the `data` before passing to `model.quant` but this goes back to a `TypeError` ```python def wrap_qat_forward_context( quant_cb, model: 'pl.LightningModule', func: Callable, trigger_condition: Optional[Union[Callable, int]] = None ) -> Callable: @functools.wraps(func) def wrapper(*data, **kwargs) -> Any: _is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer) _is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition _quant_run = trigger_condition is None or _is_func_true or _is_count_true if _quant_run: quant_cb._forward_calls += 1 data = model.quant(*data) # <-- Attempted tuple unpacking, get TypeError: forward() takes 2 positional arguments but 3 were given data = func(*data, **kwargs) if _quant_run: data = model.dequant(*data) return data return wrapper ```
non_test
qat callback doesn t work with multi input models or forward kwargs 🐛 bug the quantizationawaretraining can only handle models whose forward signature looks like forward data adding additional positional or keyword arguments breaks the forward propagation please reproduce using the boringmodel the boringmodel doesn t have the right forward signature to produce errors but i included an equivalently simple model below to reproduce python import functools import torch import torch nn as nn class simple nn module def init self super simple self init self conv nn kernel size self softmax nn softmax dim def forward self a b softmax false use positional arguments and keyword argument feat b self conv b feat a self conv a combined feat a feat b if softmax return self softmax combined else return combined property def example input array self return torch randn torch randn def wrap quantize forward context model func from quantizationawaretraining callback functools wraps func def wrapper data data model quant data data func data data model dequant data return data return wrapper works fine without quantization model simple input sample model example input array model input sample model input sample softmax true simulate the on fit start hook behavior model quant torch quantization quantstub model dequant torch quantization dequantstub model forward wrap quantize forward context model model forward model input sample model input sample softmax true the last two lines will raise the following exceptions typeerror wrapper takes positional argument but were given typeerror wrapper got an unexpected keyword argument softmax expected behavior forward prop should work fine for both the quantized and unquantized model environment cuda gpu geforce gtx with max q design available true version packages numpy pytorch debug false pytorch version pytorch lightning tqdm system os linux architecture processor python version ubuntu smp sat may utc additional context the solution to the minimal example is to make the decoration wrapper more generic python def wrap quantize forward context model func from quantizationawaretraining callback functools wraps func def wrapper data kwargs data model quant data data func data kwargs data model dequant data return data return wrapper however i notice that callbacks quantization py also includes a second decorator wrap qat forward context and i am not exactly clear on the intent behind having two separate decorators i see that wrap quantize forward context is only used in the on fit end hook any insight here borda i can make a pr for the above solution but first i tested it in an actual training by patching my local install python def wrap qat forward context quant cb model pl lightningmodule func callable trigger condition optional none callable functools wraps func def wrapper data kwargs any changes to decoration signature is func true isinstance trigger condition callable and trigger condition model trainer is count true isinstance trigger condition int and quant cb forward calls trigger condition quant run trigger condition is none or is func true or is count true if quant run quant cb forward calls data model quant data triggers attributeerror tuple object has no attribute detach data func data kwargs if quant run data model dequant data return data return wrapper so i tried also expanding the data before passing to model quant but this goes back to a typeerror python def wrap qat forward context quant cb model pl lightningmodule func callable trigger condition optional none callable functools wraps func def wrapper data kwargs any is func true isinstance trigger condition callable and trigger condition model trainer is count true isinstance trigger condition int and quant cb forward calls trigger condition quant run trigger condition is none or is func true or is count true if quant run quant cb forward calls data model quant data attempted tuple unpacking get typeerror forward takes positional arguments but were given data func data kwargs if quant run data model dequant data return data return wrapper
0
37,495
5,117,350,862
IssuesEvent
2017-01-07 15:55:16
d3athrow/vgstation13
https://api.github.com/repos/d3athrow/vgstation13
closed
95% sure killer tomatoes have been broken since botany3
Needs Moar Testing
In botany2, killer tomatoes were a random chance from tomato mutations, it went as follows 33% blue tomato 33% blood tomato 33% killer tomato Now in botany3, killer tomato is a mutation of blood tomato. However ever since it's release I have yet to get a killer tomato outside of xenobotany means (random seed happens to have killer tomato sprite) and I've never seen them grown by other botanist either. I'm pretty sure that they're broken, as I have no problem with any other mutation except this very specific one.
1.0
95% sure killer tomatoes have been broken since botany3 - In botany2, killer tomatoes were a random chance from tomato mutations, it went as follows 33% blue tomato 33% blood tomato 33% killer tomato Now in botany3, killer tomato is a mutation of blood tomato. However ever since it's release I have yet to get a killer tomato outside of xenobotany means (random seed happens to have killer tomato sprite) and I've never seen them grown by other botanist either. I'm pretty sure that they're broken, as I have no problem with any other mutation except this very specific one.
test
sure killer tomatoes have been broken since in killer tomatoes were a random chance from tomato mutations it went as follows blue tomato blood tomato killer tomato now in killer tomato is a mutation of blood tomato however ever since it s release i have yet to get a killer tomato outside of xenobotany means random seed happens to have killer tomato sprite and i ve never seen them grown by other botanist either i m pretty sure that they re broken as i have no problem with any other mutation except this very specific one
1
686,199
23,481,566,318
IssuesEvent
2022-08-17 11:03:03
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Crash returning to page with speedreader enabled
crash priority/P2 QA/Yes release-notes/exclude feature/speedreader OS/Desktop
IMPORTANT: Your crash has already been automatically reported to our crash system. Please file this bug only if you can provide more information about it. Brave Version: 1.43.58 Chromium: 104.0.5112.81 Operating System: Linux 5.18.16-200.fc36.x86_64 URL (if applicable) where crash occurred: https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html Can you reproduce this crash? No. Was reliable for a while, but currently not re-occuring. What steps will reproduce this crash? (If it's not reproducible, what were you doing just before the crash?) 1. Visit https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html with speedreader enabled 2. Follow link to doc.rust-lang.org 3. Click back button to navigate back to the blog post 4. Tab reports segv NB: Still occurs if I click 'Show original page' first. ****DO NOT CHANGE BELOW THIS LINE**** Crash ID: crash/1bfd0000-f3fb-a908-0000-000000000000
1.0
Crash returning to page with speedreader enabled - IMPORTANT: Your crash has already been automatically reported to our crash system. Please file this bug only if you can provide more information about it. Brave Version: 1.43.58 Chromium: 104.0.5112.81 Operating System: Linux 5.18.16-200.fc36.x86_64 URL (if applicable) where crash occurred: https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html Can you reproduce this crash? No. Was reliable for a while, but currently not re-occuring. What steps will reproduce this crash? (If it's not reproducible, what were you doing just before the crash?) 1. Visit https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html with speedreader enabled 2. Follow link to doc.rust-lang.org 3. Click back button to navigate back to the blog post 4. Tab reports segv NB: Still occurs if I click 'Show original page' first. ****DO NOT CHANGE BELOW THIS LINE**** Crash ID: crash/1bfd0000-f3fb-a908-0000-000000000000
non_test
crash returning to page with speedreader enabled important your crash has already been automatically reported to our crash system please file this bug only if you can provide more information about it brave version chromium operating system linux url if applicable where crash occurred can you reproduce this crash no was reliable for a while but currently not re occuring what steps will reproduce this crash if it s not reproducible what were you doing just before the crash visit with speedreader enabled follow link to doc rust lang org click back button to navigate back to the blog post tab reports segv nb still occurs if i click show original page first do not change below this line crash id crash
0
198,261
14,970,502,961
IssuesEvent
2021-01-27 19:42:48
istio/istio
https://api.github.com/repos/istio/istio
closed
Mitigate dockerhub image rate limitting
area/test and release
https://docs.docker.com/docker-hub/download-rate-limit/ CI Impact: * Base image is on docker.io, pulled each run * Kind image is on docker.io, pulled each run * buildkit image is on docker.io, pulled each run User impact: * May be limited in number of images they can pull, especially proxy image
1.0
Mitigate dockerhub image rate limitting - https://docs.docker.com/docker-hub/download-rate-limit/ CI Impact: * Base image is on docker.io, pulled each run * Kind image is on docker.io, pulled each run * buildkit image is on docker.io, pulled each run User impact: * May be limited in number of images they can pull, especially proxy image
test
mitigate dockerhub image rate limitting ci impact base image is on docker io pulled each run kind image is on docker io pulled each run buildkit image is on docker io pulled each run user impact may be limited in number of images they can pull especially proxy image
1
432,962
12,500,673,408
IssuesEvent
2020-06-01 22:55:12
ngscheurich/dice_magick
https://api.github.com/repos/ngscheurich/dice_magick
opened
Better test coverage
Priority: High Type: Maintenance
There's lots of untested code currently. We should try and get to > 90% test coverage and implement a test coverage reporter in CI, e.g. Coveralls.
1.0
Better test coverage - There's lots of untested code currently. We should try and get to > 90% test coverage and implement a test coverage reporter in CI, e.g. Coveralls.
non_test
better test coverage there s lots of untested code currently we should try and get to test coverage and implement a test coverage reporter in ci e g coveralls
0
103,224
8,883,394,590
IssuesEvent
2019-01-14 15:35:45
khartec/waltz
https://api.github.com/repos/khartec/waltz
closed
Search: support people search via email on main search and also entity search control
fixed (test & close) noteworthy
for cases where people have names with fewer characters and to speed along the search if someone has an email address to search by.
1.0
Search: support people search via email on main search and also entity search control - for cases where people have names with fewer characters and to speed along the search if someone has an email address to search by.
test
search support people search via email on main search and also entity search control for cases where people have names with fewer characters and to speed along the search if someone has an email address to search by
1
43,532
7,049,956,228
IssuesEvent
2018-01-03 01:48:45
erdl/legacy-scrape-util
https://api.github.com/repos/erdl/legacy-scrape-util
closed
Create Notebook to obtain past egauge readings in the DHHL project
prio: high stat:in progress type: documentation type: new feature
Eileen pointed out readings are still missing for DHHL egauges. We were unsure how to execute Forrest script manually to retrieve past data, so I tried using the code **locally** to see how it could be done. [A page on the wiki was added documenting the process](https://github.com/erdl/scrape-util/wiki/Manually-Running-Scrape-Util-to-Retrieve-Past-Data).
1.0
Create Notebook to obtain past egauge readings in the DHHL project - Eileen pointed out readings are still missing for DHHL egauges. We were unsure how to execute Forrest script manually to retrieve past data, so I tried using the code **locally** to see how it could be done. [A page on the wiki was added documenting the process](https://github.com/erdl/scrape-util/wiki/Manually-Running-Scrape-Util-to-Retrieve-Past-Data).
non_test
create notebook to obtain past egauge readings in the dhhl project eileen pointed out readings are still missing for dhhl egauges we were unsure how to execute forrest script manually to retrieve past data so i tried using the code locally to see how it could be done
0
486,617
14,011,910,872
IssuesEvent
2020-10-29 08:14:05
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.instagram.com - site is not usable
browser-fenix engine-gecko priority-critical
<!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/60643 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.instagram.com/ **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201026092834</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/10/0294b4c0-26f7-46c6-abb4-63d73deebac2) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.instagram.com - site is not usable - <!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/60643 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.instagram.com/ **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201026092834</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/10/0294b4c0-26f7-46c6-abb4-63d73deebac2) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
295,472
25,478,441,426
IssuesEvent
2022-11-25 16:56:57
saleor/saleor-dashboard
https://api.github.com/repos/saleor/saleor-dashboard
opened
Cypress test fail: should recalculate weight after changing shipping weight unit. TC: SALEOR_0901
tests
**Known bug for versions:** v35: false **Additional Info:** Spec: As a staff user I want to change shop default weight unit
1.0
Cypress test fail: should recalculate weight after changing shipping weight unit. TC: SALEOR_0901 - **Known bug for versions:** v35: false **Additional Info:** Spec: As a staff user I want to change shop default weight unit
test
cypress test fail should recalculate weight after changing shipping weight unit tc saleor known bug for versions false additional info spec as a staff user i want to change shop default weight unit
1
57,013
6,536,298,751
IssuesEvent
2017-08-31 17:35:57
w3c/web-platform-tests
https://api.github.com/repos/w3c/web-platform-tests
opened
Unhelpful failure reporting
testharness.js
Originally posted as https://github.com/w3c/testharness.js/issues/75 by @darobin on 27 May 2014, 14:09 UTC: > In http://www.w3c-test.org/html/browsers/the-window-object/window-properties.html the first failure I get with Firefox, for test "EventTarget method: addEventListener", is: > > ``` > assert_equals: expected function "function addEventListener() { [native code] }" but got function "function addEventListener() { [native code] }" > ```
1.0
Unhelpful failure reporting - Originally posted as https://github.com/w3c/testharness.js/issues/75 by @darobin on 27 May 2014, 14:09 UTC: > In http://www.w3c-test.org/html/browsers/the-window-object/window-properties.html the first failure I get with Firefox, for test "EventTarget method: addEventListener", is: > > ``` > assert_equals: expected function "function addEventListener() { [native code] }" but got function "function addEventListener() { [native code] }" > ```
test
unhelpful failure reporting originally posted as by darobin on may utc in the first failure i get with firefox for test eventtarget method addeventlistener is assert equals expected function function addeventlistener but got function function addeventlistener
1
120,062
12,058,407,290
IssuesEvent
2020-04-15 17:24:19
ift-gftc/doc.gdst
https://api.github.com/repos/ift-gftc/doc.gdst
closed
DOC | XML Mappings for Wild Harvest Events
documentation
## 📖 Document We need to add in XML Mappings for the Wild Harvest events.
1.0
DOC | XML Mappings for Wild Harvest Events - ## 📖 Document We need to add in XML Mappings for the Wild Harvest events.
non_test
doc xml mappings for wild harvest events 📖 document we need to add in xml mappings for the wild harvest events
0
260,122
22,593,347,116
IssuesEvent
2022-06-28 22:23:47
phetsims/energy-skate-park
https://api.github.com/repos/phetsims/energy-skate-park
closed
CT Cannot read properties of undefined (reading 'typeName')
type:automated-testing
Also in basics: ``` energy-skate-park : fuzz : built-phet-io https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone Query: fuzz&memoryLimit=1000&phetioStandalone Uncaught TypeError: Cannot read properties of undefined (reading 'typeName') TypeError: Cannot read properties of undefined (reading 'typeName') at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576) at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587) at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939) at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281) at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239) at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194) at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846) at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : fuzz : built-phet-io https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone Query: fuzz&memoryLimit=1000&phetioStandalone Uncaught TypeError: Cannot read properties of undefined (reading 'typeName') TypeError: Cannot read properties of undefined (reading 'typeName') at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576) at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587) at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939) at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281) at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239) at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194) at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846) at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655804863703%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655819123377%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-wrappers-tests : assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park&phetioDebug=true&phetioWrapperDebug=true 27 out of 28 tests passed. 1 failed. SimTests: energy-skate-park: iframe API failed: Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15) at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32) at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32 at Array.forEach (<anonymous>) at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38) at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32 at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9) Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-wrappers-tests : no-assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park 25 out of 26 tests passed. 1 failed. SimTests: energy-skate-park: iframe API failed: Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15) at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32) at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32 at Array.forEach (<anonymous>) at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38) at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32 at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9) Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ```
1.0
CT Cannot read properties of undefined (reading 'typeName') - Also in basics: ``` energy-skate-park : fuzz : built-phet-io https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone Query: fuzz&memoryLimit=1000&phetioStandalone Uncaught TypeError: Cannot read properties of undefined (reading 'typeName') TypeError: Cannot read properties of undefined (reading 'typeName') at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576) at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587) at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939) at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281) at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239) at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194) at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846) at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : fuzz : built-phet-io https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone Query: fuzz&memoryLimit=1000&phetioStandalone Uncaught TypeError: Cannot read properties of undefined (reading 'typeName') TypeError: Cannot read properties of undefined (reading 'typeName') at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576) at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587) at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939) at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180) at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281) at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239) at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194) at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846) at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655804863703%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655819123377%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-wrappers-tests : assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park&phetioDebug=true&phetioWrapperDebug=true 27 out of 28 tests passed. 1 failed. SimTests: energy-skate-park: iframe API failed: Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15) at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32) at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32 at Array.forEach (<anonymous>) at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38) at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32 at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9) Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ---------------------------------- energy-skate-park : phet-io-wrappers-tests : no-assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park 25 out of 26 tests passed. 1 failed. SimTests: energy-skate-park: iframe API failed: Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15) at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32) at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32 at Array.forEach (<anonymous>) at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38) at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32 at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9) Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: * mark the type as phetioState: false * create a custom toStateObject method in your IO Type * perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions: at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13) at assert (IOType.ts:531:14) at toStateObject (IOType.ts:267:31) at toStateObject (PhetioStateEngine.js:106:48) at getValueJSON (PhetioStateEngine.js:165:33) at Array.forEach at forEach (PhetioStateEngine.js:161:37) at getState (PhetioStateEngine.js:453:29) at initialize (phetioEngine.js:250:31) at listener (TinyEmitter.ts:94:8) id: Bayes Puppeteer Snapshot from 6/21/2022, 3:20:44 AM ```
test
ct cannot read properties of undefined reading typename also in basics energy skate park fuzz built phet io query fuzz memorylimit phetiostandalone uncaught typeerror cannot read properties of undefined reading typename typeerror cannot read properties of undefined reading typename at h at object getdata at v phetiostartevent at v notifylisteners at v set at v set at v set value at at n emit at m notifylisteners id bayes puppeteer snapshot from am energy skate park fuzz built phet io query fuzz memorylimit phetiostandalone uncaught typeerror cannot read properties of undefined reading typename typeerror cannot read properties of undefined reading typename at h at object getdata at v phetiostartevent at v notifylisteners at v set at v set at v set value at at n emit at m notifylisteners id bayes puppeteer snapshot from am energy skate park phet io fuzz unbuilt query ea brand phet io phetiostandalone fuzz memorylimit uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am energy skate park phet io fuzz unbuilt query ea brand phet io phetiostandalone fuzz memorylimit uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am energy skate park phet io wrappers tests assert out of tests passed failed simtests energy skate park iframe api failed uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor at window assertions assertfunction at object tostateobject at iotype tostateobject at phetiostateengine getvaluejson at at array foreach at phetiostateengine getstate at phetiostateengine initialize at at tinyproperty emit uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructorerror assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am energy skate park phet io wrappers tests no assert out of tests passed failed simtests energy skate park iframe api failed uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor at window assertions assertfunction at object tostateobject at iotype tostateobject at phetiostateengine getvaluejson at at array foreach at phetiostateengine getstate at phetiostateengine initialize at at tinyproperty emit uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructorerror assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am
1
55,103
6,426,548,207
IssuesEvent
2017-08-09 17:42:03
tcpd/surf
https://api.github.com/repos/tcpd/surf
closed
Not able to read back a file generated by surf_worksheet_generator.R in Surf
Fixed. Waiting for Test P1 - High Priority
All the merges are lost. Needs to be fixed asap.
1.0
Not able to read back a file generated by surf_worksheet_generator.R in Surf - All the merges are lost. Needs to be fixed asap.
test
not able to read back a file generated by surf worksheet generator r in surf all the merges are lost needs to be fixed asap
1
17,925
4,215,963,298
IssuesEvent
2016-06-30 07:24:18
nemerosa/ontrack
https://api.github.com/repos/nemerosa/ontrack
opened
Generation of the Ontrack documentation
documentation
* [ ] Embed the Ontrack documentation in the code * [ ] Generate both single page and PDF documentation * [ ] Publish PDF documentation in GitHub release * [ ] Publish HTML documentation in Ontrack GitHub pages * [ ] Migrate the Wiki documentation to the embedded documentation
1.0
Generation of the Ontrack documentation - * [ ] Embed the Ontrack documentation in the code * [ ] Generate both single page and PDF documentation * [ ] Publish PDF documentation in GitHub release * [ ] Publish HTML documentation in Ontrack GitHub pages * [ ] Migrate the Wiki documentation to the embedded documentation
non_test
generation of the ontrack documentation embed the ontrack documentation in the code generate both single page and pdf documentation publish pdf documentation in github release publish html documentation in ontrack github pages migrate the wiki documentation to the embedded documentation
0
33,281
7,694,311,749
IssuesEvent
2018-05-18 08:19:15
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
JIT: assert when jitting Sse2.ConvertToInt32WithTruncation
area-CodeGen bug
```C# using System; using System.Runtime.Intrinsics; using System.Runtime.Intrinsics.X86; class X { public static int Main() { Vector128<double> x = new Vector128<double>(); int y = Sse2.ConvertToInt32WithTruncation(x); return y; } } ``` Running this with a CHECKED build hits an assert: ``` Assert failure(PID 25028 [0x000061c4], Thread: 14596 [0x3904]): Assertion failed 'unreached' in 'X:Main():int' (IL size 23) File: d:\repos\coreclr\src\jit\hwintrinsiccodegenxarch.cpp Line: 1103 Image: D:\repos\coreclr\bin\tests\Windows_NT.x64.Checked\Tests\Core_Root\CoreRun.exe ```
1.0
JIT: assert when jitting Sse2.ConvertToInt32WithTruncation - ```C# using System; using System.Runtime.Intrinsics; using System.Runtime.Intrinsics.X86; class X { public static int Main() { Vector128<double> x = new Vector128<double>(); int y = Sse2.ConvertToInt32WithTruncation(x); return y; } } ``` Running this with a CHECKED build hits an assert: ``` Assert failure(PID 25028 [0x000061c4], Thread: 14596 [0x3904]): Assertion failed 'unreached' in 'X:Main():int' (IL size 23) File: d:\repos\coreclr\src\jit\hwintrinsiccodegenxarch.cpp Line: 1103 Image: D:\repos\coreclr\bin\tests\Windows_NT.x64.Checked\Tests\Core_Root\CoreRun.exe ```
non_test
jit assert when jitting c using system using system runtime intrinsics using system runtime intrinsics class x public static int main x new int y x return y running this with a checked build hits an assert assert failure pid thread assertion failed unreached in x main int il size file d repos coreclr src jit hwintrinsiccodegenxarch cpp line image d repos coreclr bin tests windows nt checked tests core root corerun exe
0
333,664
29,801,340,309
IssuesEvent
2023-06-16 08:20:35
bitsstdcheee/luogu-status
https://api.github.com/repos/bitsstdcheee/luogu-status
closed
🛑 Luogu Contest is down
status luogu-contest
In [`74802cc`](https://github.com/bitsstdcheee/luogu-status/commit/74802cc4854ee533b23a229bcf3390051a6653c4 ), Luogu Contest (https://www.luogu.com.cn/contest/list) was **down**: - HTTP code: 403 - Response time: 12 ms
1.0
🛑 Luogu Contest is down - In [`74802cc`](https://github.com/bitsstdcheee/luogu-status/commit/74802cc4854ee533b23a229bcf3390051a6653c4 ), Luogu Contest (https://www.luogu.com.cn/contest/list) was **down**: - HTTP code: 403 - Response time: 12 ms
test
🛑 luogu contest is down in luogu contest was down http code response time ms
1
31,830
26,170,631,530
IssuesEvent
2023-01-01 21:48:59
ManimCommunity/manim
https://api.github.com/repos/ManimCommunity/manim
closed
Update Poetry Installer to New Version
infrastructure
The old poetry installer has been deprecated for a while now, but we still use it. In my short (but not perfect) testing, it appears to me that the old installer will have issues on 3.10 (Which as of the creation of this issue is 6 days from launch ) This will require large changes to our CI that I have struggled to do on my own, hence the issue.
1.0
Update Poetry Installer to New Version - The old poetry installer has been deprecated for a while now, but we still use it. In my short (but not perfect) testing, it appears to me that the old installer will have issues on 3.10 (Which as of the creation of this issue is 6 days from launch ) This will require large changes to our CI that I have struggled to do on my own, hence the issue.
non_test
update poetry installer to new version the old poetry installer has been deprecated for a while now but we still use it in my short but not perfect testing it appears to me that the old installer will have issues on which as of the creation of this issue is days from launch this will require large changes to our ci that i have struggled to do on my own hence the issue
0
142,363
11,470,124,021
IssuesEvent
2020-02-09 00:16:46
namhyung/uftrace
https://api.github.com/repos/namhyung/uftrace
opened
clang compiled uftrace test result
tests
I've compiled uftrace with clang-9.0.1 and run tests. The result is as follows: ``` Test case pg finstrument-fu ------------------------: O0 O1 O2 O3 Os O0 O1 O2 O3 Os 015 longjmp : OK NG NG NG NG OK OK OK OK OK 051 return : NG NG NG NG NG NG OK OK OK OK 052 nested_func : BI BI BI BI BI BI BI BI BI BI 062 arg_char : OK NG NG NG NG SK SK SK SK SK 082 arg_many : NG OK OK OK OK SK SK SK SK SK 083 arg_float : NG NG NG NG NG SK SK SK SK SK 084 arg_mixed : NG NG NG NG NG SK SK SK SK SK 085 arg_reg : NG NG NG NG NG SK SK SK SK SK 087 arg_variadic : NG NG NG NG NG SK SK SK SK SK 090 report_recursive : OK NG NG NG NG OK OK OK OK OK 124 exception : OK OK OK OK OK NG NG NG NG NG 125 report_range : OK NG OK OK OK OK OK OK OK OK 136 dynamic : BI BI BI BI BI BI BI BI BI BI 151 recv_runcmd : NG NG NG NG NG NG NG NG NG NG 162 pltbind_now_pie : BI BI BI BI BI BI BI BI BI BI 181 graph_full : NG NG NG NG NG NG NG NG NG NG 182 thread_exit : SG OK OK NG OK SG OK NG NG NG 184 arg_enum : NG NG NG NG NG NG NG NG NG NG 185 exception2 : OK OK OK OK OK NG NG NG NG NG 186 exception3 : OK OK OK OK OK NG NG NG NG NG 191 posix_spawn : NG OK OK OK OK NG OK OK OK OK 198 lib_arg_float : NG NG NG NG NG SK SK SK SK SK 201 arg_dwarf1 : NG NG NG NG NG SK SK SK SK SK 202 arg_dwarf2 : NG OK OK OK OK SK SK SK SK SK 203 arg_dwarf3 : NG NG NG NG NG SK SK SK SK SK 222 external_data : OK OK OK OK OK OK OK NG OK OK 225 dynamic_size : NG OK OK OK NG NG OK OK OK NG 228 read_pmu_cycle3 : OK NG NG NG NG OK OK NG OK OK 231 arg_bound : NG OK OK OK OK SK SK SK SK SK 232 dynamic_unpatch : BI BI BI BI BI BI BI BI BI BI ```
1.0
clang compiled uftrace test result - I've compiled uftrace with clang-9.0.1 and run tests. The result is as follows: ``` Test case pg finstrument-fu ------------------------: O0 O1 O2 O3 Os O0 O1 O2 O3 Os 015 longjmp : OK NG NG NG NG OK OK OK OK OK 051 return : NG NG NG NG NG NG OK OK OK OK 052 nested_func : BI BI BI BI BI BI BI BI BI BI 062 arg_char : OK NG NG NG NG SK SK SK SK SK 082 arg_many : NG OK OK OK OK SK SK SK SK SK 083 arg_float : NG NG NG NG NG SK SK SK SK SK 084 arg_mixed : NG NG NG NG NG SK SK SK SK SK 085 arg_reg : NG NG NG NG NG SK SK SK SK SK 087 arg_variadic : NG NG NG NG NG SK SK SK SK SK 090 report_recursive : OK NG NG NG NG OK OK OK OK OK 124 exception : OK OK OK OK OK NG NG NG NG NG 125 report_range : OK NG OK OK OK OK OK OK OK OK 136 dynamic : BI BI BI BI BI BI BI BI BI BI 151 recv_runcmd : NG NG NG NG NG NG NG NG NG NG 162 pltbind_now_pie : BI BI BI BI BI BI BI BI BI BI 181 graph_full : NG NG NG NG NG NG NG NG NG NG 182 thread_exit : SG OK OK NG OK SG OK NG NG NG 184 arg_enum : NG NG NG NG NG NG NG NG NG NG 185 exception2 : OK OK OK OK OK NG NG NG NG NG 186 exception3 : OK OK OK OK OK NG NG NG NG NG 191 posix_spawn : NG OK OK OK OK NG OK OK OK OK 198 lib_arg_float : NG NG NG NG NG SK SK SK SK SK 201 arg_dwarf1 : NG NG NG NG NG SK SK SK SK SK 202 arg_dwarf2 : NG OK OK OK OK SK SK SK SK SK 203 arg_dwarf3 : NG NG NG NG NG SK SK SK SK SK 222 external_data : OK OK OK OK OK OK OK NG OK OK 225 dynamic_size : NG OK OK OK NG NG OK OK OK NG 228 read_pmu_cycle3 : OK NG NG NG NG OK OK NG OK OK 231 arg_bound : NG OK OK OK OK SK SK SK SK SK 232 dynamic_unpatch : BI BI BI BI BI BI BI BI BI BI ```
test
clang compiled uftrace test result i ve compiled uftrace with clang and run tests the result is as follows test case pg finstrument fu os os longjmp ok ng ng ng ng ok ok ok ok ok return ng ng ng ng ng ng ok ok ok ok nested func bi bi bi bi bi bi bi bi bi bi arg char ok ng ng ng ng sk sk sk sk sk arg many ng ok ok ok ok sk sk sk sk sk arg float ng ng ng ng ng sk sk sk sk sk arg mixed ng ng ng ng ng sk sk sk sk sk arg reg ng ng ng ng ng sk sk sk sk sk arg variadic ng ng ng ng ng sk sk sk sk sk report recursive ok ng ng ng ng ok ok ok ok ok exception ok ok ok ok ok ng ng ng ng ng report range ok ng ok ok ok ok ok ok ok ok dynamic bi bi bi bi bi bi bi bi bi bi recv runcmd ng ng ng ng ng ng ng ng ng ng pltbind now pie bi bi bi bi bi bi bi bi bi bi graph full ng ng ng ng ng ng ng ng ng ng thread exit sg ok ok ng ok sg ok ng ng ng arg enum ng ng ng ng ng ng ng ng ng ng ok ok ok ok ok ng ng ng ng ng ok ok ok ok ok ng ng ng ng ng posix spawn ng ok ok ok ok ng ok ok ok ok lib arg float ng ng ng ng ng sk sk sk sk sk arg ng ng ng ng ng sk sk sk sk sk arg ng ok ok ok ok sk sk sk sk sk arg ng ng ng ng ng sk sk sk sk sk external data ok ok ok ok ok ok ok ng ok ok dynamic size ng ok ok ok ng ng ok ok ok ng read pmu ok ng ng ng ng ok ok ng ok ok arg bound ng ok ok ok ok sk sk sk sk sk dynamic unpatch bi bi bi bi bi bi bi bi bi bi
1
636,120
20,592,545,103
IssuesEvent
2022-03-05 02:25:37
apcountryman/picolibrary-microchip-megaavr
https://api.github.com/repos/apcountryman/picolibrary-microchip-megaavr
closed
Add Microchip megaAVR I2C controller interactive testing skeleton
priority-normal status-awaiting_review type-feature
Add Microchip megaAVR I2C controller (`::picolibrary::Microchip::megaAVR::Controller`) interactive testing skeleton. - [x] Add `test/interactive/picolibrary/microchip/megaavr/i2c/controller/CMakeLists.txt` - [x] Add the `test/interactive/picolibrary/microchip/megaavr/i2c/controller/` directory to the build
1.0
Add Microchip megaAVR I2C controller interactive testing skeleton - Add Microchip megaAVR I2C controller (`::picolibrary::Microchip::megaAVR::Controller`) interactive testing skeleton. - [x] Add `test/interactive/picolibrary/microchip/megaavr/i2c/controller/CMakeLists.txt` - [x] Add the `test/interactive/picolibrary/microchip/megaavr/i2c/controller/` directory to the build
non_test
add microchip megaavr controller interactive testing skeleton add microchip megaavr controller picolibrary microchip megaavr controller interactive testing skeleton add test interactive picolibrary microchip megaavr controller cmakelists txt add the test interactive picolibrary microchip megaavr controller directory to the build
0
44,858
5,657,410,782
IssuesEvent
2017-04-10 07:00:20
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Add CI tests for geo-centroid
:Tilemap :Visualizations test v5.4.0
the selenium tests should test the new geo-centroid feature, which is currently missing.
1.0
Add CI tests for geo-centroid - the selenium tests should test the new geo-centroid feature, which is currently missing.
test
add ci tests for geo centroid the selenium tests should test the new geo centroid feature which is currently missing
1
274,488
23,842,521,910
IssuesEvent
2022-09-06 11:29:02
stores-cedcommerce/Lucy-Store-Design
https://api.github.com/repos/stores-cedcommerce/Lucy-Store-Design
closed
The Refund policy title needed to be change.
Footer section Desktop Ready to test Content Type / typo fixed
**Actual result:** The Refund policy title needed to be change. ![image](https://user-images.githubusercontent.com/102131636/187645204-7473ea3e-20cf-4bf5-ab8a-7642c07d0739.png) **Expected result:** The title needed to be updated like Returns and Refund Policy and the titles inside the page needed to be updated.
1.0
The Refund policy title needed to be change. - **Actual result:** The Refund policy title needed to be change. ![image](https://user-images.githubusercontent.com/102131636/187645204-7473ea3e-20cf-4bf5-ab8a-7642c07d0739.png) **Expected result:** The title needed to be updated like Returns and Refund Policy and the titles inside the page needed to be updated.
test
the refund policy title needed to be change actual result the refund policy title needed to be change expected result the title needed to be updated like returns and refund policy and the titles inside the page needed to be updated
1
348,079
31,465,781,990
IssuesEvent
2023-08-30 01:48:00
hoprnet/hoprnet
https://api.github.com/repos/hoprnet/hoprnet
closed
loadtesting HOPRd nodes
bug testing stale
Related to https://github.com/Rpc-h/RPCh/pull/403 Running three types of load tests against RPCh + HOPRd nodes. I've confirmed that the bottleneck is reproducible directly via using the HOPRd API so there isn't a bottleneck between RPCh <> HOPRd nodes (meaning the `rpc-server`). Using two dedicated nodes, one acting as an entry and another as an exit. image: `gcr.io/hoprassociation/hoprd:1682855128` environment: `rpch staging` The tests are meant to test `deliverability` and do not concern themselves with response time by intention. Summary of the results of a single run: - constant load test (run various load tests for 10 seconds each) - 10 reqs in 10 seconds (1 req/s) = 100% delivery - 100 reqs in 10 seconds (10 req/s) = 88% delivery - 1000 reqs in 10 seconds (100 req/s) = 22% delivery - burst load test (run various burst tests for 1 second each) - 10 reqs in 1 second = 100% delivery - 50 reqs in 1 second = 94% delivery - 100 reqs in 1 second = 98% delivery Additional context: Why is this a problem for RPCh? You can see that already with 10 reqs per second (which can easily be achieved by 1 RPCh user) we start experiencing large delivery losses. Additionally, as soon as the node's connected users are more than 1, the delivery losses keep increasing.
1.0
loadtesting HOPRd nodes - Related to https://github.com/Rpc-h/RPCh/pull/403 Running three types of load tests against RPCh + HOPRd nodes. I've confirmed that the bottleneck is reproducible directly via using the HOPRd API so there isn't a bottleneck between RPCh <> HOPRd nodes (meaning the `rpc-server`). Using two dedicated nodes, one acting as an entry and another as an exit. image: `gcr.io/hoprassociation/hoprd:1682855128` environment: `rpch staging` The tests are meant to test `deliverability` and do not concern themselves with response time by intention. Summary of the results of a single run: - constant load test (run various load tests for 10 seconds each) - 10 reqs in 10 seconds (1 req/s) = 100% delivery - 100 reqs in 10 seconds (10 req/s) = 88% delivery - 1000 reqs in 10 seconds (100 req/s) = 22% delivery - burst load test (run various burst tests for 1 second each) - 10 reqs in 1 second = 100% delivery - 50 reqs in 1 second = 94% delivery - 100 reqs in 1 second = 98% delivery Additional context: Why is this a problem for RPCh? You can see that already with 10 reqs per second (which can easily be achieved by 1 RPCh user) we start experiencing large delivery losses. Additionally, as soon as the node's connected users are more than 1, the delivery losses keep increasing.
test
loadtesting hoprd nodes related to running three types of load tests against rpch hoprd nodes i ve confirmed that the bottleneck is reproducible directly via using the hoprd api so there isn t a bottleneck between rpch hoprd nodes meaning the rpc server using two dedicated nodes one acting as an entry and another as an exit image gcr io hoprassociation hoprd environment rpch staging the tests are meant to test deliverability and do not concern themselves with response time by intention summary of the results of a single run constant load test run various load tests for seconds each reqs in seconds req s delivery reqs in seconds req s delivery reqs in seconds req s delivery burst load test run various burst tests for second each reqs in second delivery reqs in second delivery reqs in second delivery additional context why is this a problem for rpch you can see that already with reqs per second which can easily be achieved by rpch user we start experiencing large delivery losses additionally as soon as the node s connected users are more than the delivery losses keep increasing
1
340,482
30,518,856,204
IssuesEvent
2023-07-19 06:30:32
guram-asanidze-1-btu-edu-ge/shualeduri
https://api.github.com/repos/guram-asanidze-1-btu-edu-ge/shualeduri
opened
435d763 failed unit and formatting tests.
ci-pytest ci-black
Automatically generated message 435d7632962fde65708858b27fbea1ae9d0205b9 failed unit and formatting tests. Pytest report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/pytest.html Black report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/black.html
1.0
435d763 failed unit and formatting tests. - Automatically generated message 435d7632962fde65708858b27fbea1ae9d0205b9 failed unit and formatting tests. Pytest report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/pytest.html Black report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/black.html
test
failed unit and formatting tests automatically generated message failed unit and formatting tests pytest report black report
1
223,479
17,602,475,289
IssuesEvent
2021-08-17 13:28:26
Gemini-team/Gemini
https://api.github.com/repos/Gemini-team/Gemini
closed
Unit testing in Unity
enhancement testing
Until now there has only been manual testing of the code base by running the simulator and verifying that things works as intended. This will not cut it going further, and the first step in the testing pipeline will be to get Unit testing in the Unity project up and running.
1.0
Unit testing in Unity - Until now there has only been manual testing of the code base by running the simulator and verifying that things works as intended. This will not cut it going further, and the first step in the testing pipeline will be to get Unit testing in the Unity project up and running.
test
unit testing in unity until now there has only been manual testing of the code base by running the simulator and verifying that things works as intended this will not cut it going further and the first step in the testing pipeline will be to get unit testing in the unity project up and running
1
25,693
4,165,930,802
IssuesEvent
2016-06-19 20:38:24
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Graphs only work if you are on first tab in Host and Container View
area/container area/host area/ui kind/bug status/resolved status/to-test
**Rancher Version:** master 3/16 **Steps to Reproduce:** 1. Go to Host View or Container View 2. Select Volume tab in container view or Port tab in host view **Results:** Screen noticeably slow down and the graph stop showing data **Expected:** show data
1.0
Graphs only work if you are on first tab in Host and Container View - **Rancher Version:** master 3/16 **Steps to Reproduce:** 1. Go to Host View or Container View 2. Select Volume tab in container view or Port tab in host view **Results:** Screen noticeably slow down and the graph stop showing data **Expected:** show data
test
graphs only work if you are on first tab in host and container view rancher version master steps to reproduce go to host view or container view select volume tab in container view or port tab in host view results screen noticeably slow down and the graph stop showing data expected show data
1
317,827
27,270,490,381
IssuesEvent
2023-02-22 21:50:30
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
validate-modules for plugins: for callback plugins, type vs. callback_type
test support:core bug has_pr affects_2.15
##### SUMMARY This is not a bug, but a discussion issue for one aspect of #71734 ([WIP] Extend validate-modules to also validate plugins). It provides details on one of the points in https://github.com/ansible/community/issues/560#issuecomment-691529907. Callback plugins have a type. In the plugin class, it can be set as the class attribute `CALLBACK_TYPE`. For documentation, there are two versions: 1. `callback_type: <type>` https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/tree.py#L10 2. `type: <type>` https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/default.py#L10 In ansible/ansible, `type` is much more common (there's only one callback using `callback_type`). In community.general, 1/3 of the callback plugins uses `callback_type`. I would suggest that we use only one of these two. My personal suggestion would be `callback_type`, since it is close to the `CALLBACK_TYPE` attribute of the plugin class. On the other hand, @abadger chose `type` in the antsibull-docs schemas (https://github.com/ansible-community/antsibull/blob/main/antsibull/schemas/callback.py#L23). ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME test/lib/ansible_test/_data/sanity/validate-modules/validate_modules/main.py ##### ANSIBLE VERSION 2.11 bot_skip
1.0
validate-modules for plugins: for callback plugins, type vs. callback_type - ##### SUMMARY This is not a bug, but a discussion issue for one aspect of #71734 ([WIP] Extend validate-modules to also validate plugins). It provides details on one of the points in https://github.com/ansible/community/issues/560#issuecomment-691529907. Callback plugins have a type. In the plugin class, it can be set as the class attribute `CALLBACK_TYPE`. For documentation, there are two versions: 1. `callback_type: <type>` https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/tree.py#L10 2. `type: <type>` https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/default.py#L10 In ansible/ansible, `type` is much more common (there's only one callback using `callback_type`). In community.general, 1/3 of the callback plugins uses `callback_type`. I would suggest that we use only one of these two. My personal suggestion would be `callback_type`, since it is close to the `CALLBACK_TYPE` attribute of the plugin class. On the other hand, @abadger chose `type` in the antsibull-docs schemas (https://github.com/ansible-community/antsibull/blob/main/antsibull/schemas/callback.py#L23). ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME test/lib/ansible_test/_data/sanity/validate-modules/validate_modules/main.py ##### ANSIBLE VERSION 2.11 bot_skip
test
validate modules for plugins for callback plugins type vs callback type summary this is not a bug but a discussion issue for one aspect of extend validate modules to also validate plugins it provides details on one of the points in callback plugins have a type in the plugin class it can be set as the class attribute callback type for documentation there are two versions callback type type in ansible ansible type is much more common there s only one callback using callback type in community general of the callback plugins uses callback type i would suggest that we use only one of these two my personal suggestion would be callback type since it is close to the callback type attribute of the plugin class on the other hand abadger chose type in the antsibull docs schemas issue type bug report component name test lib ansible test data sanity validate modules validate modules main py ansible version bot skip
1
184,296
14,285,928,407
IssuesEvent
2020-11-23 14:32:46
photoprism/photoprism
https://api.github.com/repos/photoprism/photoprism
closed
Physically moved files cannot be indexed again
bug important please-test
I'm encountering an issue with physically moving photos around where Photoprism thinks the moved photos are duplicates, even though they've been removed from their original position. My workflow is as follows: * Copy all files from the SD card to an 'unsorted' directory * Re-index so that they are accessible in Photoprism * Edit them later and move them to the appropriate places * Re-index the library The problem is that in the 2nd reindexing: the photos appear to not yet have been removed from the database when it scans the new directory. It therefore tags the moved photos as duplicates. By the end of the re-indexing the photos are no longer found at their original locations, however they could not be scanned again either. I've tried various methods including a complete re-scan of both the original directory and the new directory but PP doesn't recognise the new photo. My last resort would be to edit the database manually but it is not safe and inconvenient to do so every time.. The expectation is that Photoprism would be able to detect moved files with a single re-indexing. It seemed the sidecar files might be a reason for this as they are not automatically removed when the original files are removed. I've also tried removing the sidecar completely from both directories and a clean re-indexing but that also didn't help. This is an example log: ``` 2020-10-25 08:27:32 INFO Indexing completed in 21 s 2020-10-25 08:27:31 INFO purge: searching index for hidden media files 2020-10-25 08:27:17 INFO index: skipped main raw file 2018/07/7R306756.ARW 2020-10-25 08:27:10 INFO index: no .ppignore file found 2020-10-25 08:27:10 INFO Indexing originals... 2020-10-25 08:27:10 INFO Settings saved ``` The RAW file, along with the JPG and XMP files are in the right location. No *.json files for them are present, however the json file is recreated after a normal re-indexing. It appears that it was pulled from the database and put back to the filesystem?. To reproduce this issue: * Add a file to PP by indexing * Move the file elsewhere. The new directory might need to be named/placed in such a way that it is scanned before the old one in the directory hierarchy. In my case the old directory is a sub-directory of the new one. * Re-index As a workaround, the file could be reindexed if its entry is removed from the database in the "files" set.
1.0
Physically moved files cannot be indexed again - I'm encountering an issue with physically moving photos around where Photoprism thinks the moved photos are duplicates, even though they've been removed from their original position. My workflow is as follows: * Copy all files from the SD card to an 'unsorted' directory * Re-index so that they are accessible in Photoprism * Edit them later and move them to the appropriate places * Re-index the library The problem is that in the 2nd reindexing: the photos appear to not yet have been removed from the database when it scans the new directory. It therefore tags the moved photos as duplicates. By the end of the re-indexing the photos are no longer found at their original locations, however they could not be scanned again either. I've tried various methods including a complete re-scan of both the original directory and the new directory but PP doesn't recognise the new photo. My last resort would be to edit the database manually but it is not safe and inconvenient to do so every time.. The expectation is that Photoprism would be able to detect moved files with a single re-indexing. It seemed the sidecar files might be a reason for this as they are not automatically removed when the original files are removed. I've also tried removing the sidecar completely from both directories and a clean re-indexing but that also didn't help. This is an example log: ``` 2020-10-25 08:27:32 INFO Indexing completed in 21 s 2020-10-25 08:27:31 INFO purge: searching index for hidden media files 2020-10-25 08:27:17 INFO index: skipped main raw file 2018/07/7R306756.ARW 2020-10-25 08:27:10 INFO index: no .ppignore file found 2020-10-25 08:27:10 INFO Indexing originals... 2020-10-25 08:27:10 INFO Settings saved ``` The RAW file, along with the JPG and XMP files are in the right location. No *.json files for them are present, however the json file is recreated after a normal re-indexing. It appears that it was pulled from the database and put back to the filesystem?. To reproduce this issue: * Add a file to PP by indexing * Move the file elsewhere. The new directory might need to be named/placed in such a way that it is scanned before the old one in the directory hierarchy. In my case the old directory is a sub-directory of the new one. * Re-index As a workaround, the file could be reindexed if its entry is removed from the database in the "files" set.
test
physically moved files cannot be indexed again i m encountering an issue with physically moving photos around where photoprism thinks the moved photos are duplicates even though they ve been removed from their original position my workflow is as follows copy all files from the sd card to an unsorted directory re index so that they are accessible in photoprism edit them later and move them to the appropriate places re index the library the problem is that in the reindexing the photos appear to not yet have been removed from the database when it scans the new directory it therefore tags the moved photos as duplicates by the end of the re indexing the photos are no longer found at their original locations however they could not be scanned again either i ve tried various methods including a complete re scan of both the original directory and the new directory but pp doesn t recognise the new photo my last resort would be to edit the database manually but it is not safe and inconvenient to do so every time the expectation is that photoprism would be able to detect moved files with a single re indexing it seemed the sidecar files might be a reason for this as they are not automatically removed when the original files are removed i ve also tried removing the sidecar completely from both directories and a clean re indexing but that also didn t help this is an example log info indexing completed in s info purge searching index for hidden media files info index skipped main raw file arw info index no ppignore file found info indexing originals info settings saved the raw file along with the jpg and xmp files are in the right location no json files for them are present however the json file is recreated after a normal re indexing it appears that it was pulled from the database and put back to the filesystem to reproduce this issue add a file to pp by indexing move the file elsewhere the new directory might need to be named placed in such a way that it is scanned before the old one in the directory hierarchy in my case the old directory is a sub directory of the new one re index as a workaround the file could be reindexed if its entry is removed from the database in the files set
1
80,581
10,193,503,894
IssuesEvent
2019-08-12 13:48:49
catboost/catboost
https://api.github.com/repos/catboost/catboost
closed
wrong CatBoost Viewer installation instractions
bug documentation
Problem: {your page https://catboost.ai/docs/concepts/catboostviewer-installation.html#catboostviewer-installation sugests to click on npm but this references to broken link https://www.npmjs.com/package/nmp/tutorial. So how to install needed soft for CatBoost Viewer for use for Python , if possble at all} catboost version: {latest} Operating System: {Windows} CPU: {i7} # GPU: {GPU}
1.0
wrong CatBoost Viewer installation instractions - Problem: {your page https://catboost.ai/docs/concepts/catboostviewer-installation.html#catboostviewer-installation sugests to click on npm but this references to broken link https://www.npmjs.com/package/nmp/tutorial. So how to install needed soft for CatBoost Viewer for use for Python , if possble at all} catboost version: {latest} Operating System: {Windows} CPU: {i7} # GPU: {GPU}
non_test
wrong catboost viewer installation instractions problem your page sugests to click on npm but this references to broken link so how to install needed soft for catboost viewer for use for python if possble at all catboost version latest operating system windows cpu gpu gpu
0
166,546
12,961,569,887
IssuesEvent
2020-07-20 15:51:20
TBS-EACPD/infobase
https://api.github.com/repos/TBS-EACPD/infobase
closed
Add method of getting in touch to the report a problem widget
dead? enhancement user testing
There is no option in the Report a Problem Widget for users to receive an answer to potential inquiries they have about trouble with the site. There is an opportunity to use the InfoBase feedback inbox as a means of direct contact, once we reclaim access to that inbox.
1.0
Add method of getting in touch to the report a problem widget - There is no option in the Report a Problem Widget for users to receive an answer to potential inquiries they have about trouble with the site. There is an opportunity to use the InfoBase feedback inbox as a means of direct contact, once we reclaim access to that inbox.
test
add method of getting in touch to the report a problem widget there is no option in the report a problem widget for users to receive an answer to potential inquiries they have about trouble with the site there is an opportunity to use the infobase feedback inbox as a means of direct contact once we reclaim access to that inbox
1
41,483
5,358,637,780
IssuesEvent
2017-02-20 23:01:17
awslabs/s2n
https://api.github.com/repos/awslabs/s2n
closed
Add Test Integration with "BoGo" BoringSSL's Test Suite
type/test
BoringSSL has a test suite runner named "BoGo" that can run against other TLS Implementations which gets around 75% line coverage against BoringSSL. It is currently in an alpha stage for non-BoringSSL implementations, but it looks like they want to get it into a state where other implementations can use it too. I think it'd be interesting to get BoGo tests into a state where we can locally run BoGo through a "make bogo-test" or similar to see what, if any, BoGo tests s2n fails. Once we have it in a state where we can run it manually, we can work on opting out of specific tests for features that s2n doesn't implement and work on bringing our pass rate up to 100%. Links: - https://www.ietf.org/mail-archive/web/tls/current/msg20793.html - https://boringssl.googlesource.com/boringssl/+/master/ssl/test/ - https://boringssl.googlesource.com/boringssl/+/master/ssl/test/PORTING.md - https://github.com/google/openssl-tests
1.0
Add Test Integration with "BoGo" BoringSSL's Test Suite - BoringSSL has a test suite runner named "BoGo" that can run against other TLS Implementations which gets around 75% line coverage against BoringSSL. It is currently in an alpha stage for non-BoringSSL implementations, but it looks like they want to get it into a state where other implementations can use it too. I think it'd be interesting to get BoGo tests into a state where we can locally run BoGo through a "make bogo-test" or similar to see what, if any, BoGo tests s2n fails. Once we have it in a state where we can run it manually, we can work on opting out of specific tests for features that s2n doesn't implement and work on bringing our pass rate up to 100%. Links: - https://www.ietf.org/mail-archive/web/tls/current/msg20793.html - https://boringssl.googlesource.com/boringssl/+/master/ssl/test/ - https://boringssl.googlesource.com/boringssl/+/master/ssl/test/PORTING.md - https://github.com/google/openssl-tests
test
add test integration with bogo boringssl s test suite boringssl has a test suite runner named bogo that can run against other tls implementations which gets around line coverage against boringssl it is currently in an alpha stage for non boringssl implementations but it looks like they want to get it into a state where other implementations can use it too i think it d be interesting to get bogo tests into a state where we can locally run bogo through a make bogo test or similar to see what if any bogo tests fails once we have it in a state where we can run it manually we can work on opting out of specific tests for features that doesn t implement and work on bringing our pass rate up to links
1
61,188
6,726,962,886
IssuesEvent
2017-10-17 11:56:17
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
app-linux-img-converter v1.2.3 (r3.2)
r3.2-fc26-testing
Update of app-linux-img-converter to v1.2.3 for Qubes r3.2, see comments below for details. Built from: https://github.com/QubesOS/qubes-app-linux-img-converter/commit/851e4f026fcc9a45aa52e181360300640215064d [Changes since previous version](https://github.com/QubesOS/qubes-app-linux-img-converter/compare/v1.2.2...v1.2.3): QubesOS/qubes-app-linux-img-converter@851e4f0 version 1.2.3 QubesOS/qubes-app-linux-img-converter@004bc7b Removed .travis.yml bootstrap fix QubesOS/qubes-app-linux-img-converter@1d757ce Updated manpage to have correct binary name Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current repo` (available 7 days from now) * `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
1.0
app-linux-img-converter v1.2.3 (r3.2) - Update of app-linux-img-converter to v1.2.3 for Qubes r3.2, see comments below for details. Built from: https://github.com/QubesOS/qubes-app-linux-img-converter/commit/851e4f026fcc9a45aa52e181360300640215064d [Changes since previous version](https://github.com/QubesOS/qubes-app-linux-img-converter/compare/v1.2.2...v1.2.3): QubesOS/qubes-app-linux-img-converter@851e4f0 version 1.2.3 QubesOS/qubes-app-linux-img-converter@004bc7b Removed .travis.yml bootstrap fix QubesOS/qubes-app-linux-img-converter@1d757ce Updated manpage to have correct binary name Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current repo` (available 7 days from now) * `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
test
app linux img converter update of app linux img converter to for qubes see comments below for details built from qubesos qubes app linux img converter version qubesos qubes app linux img converter removed travis yml bootstrap fix qubesos qubes app linux img converter updated manpage to have correct binary name referenced issues if you re release manager you can issue gpg inline signed command upload app linux img converter current repo available days from now upload app linux img converter current dists repo you can choose subset of distributions like vm vm available days from now upload app linux img converter security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
1
612,882
19,058,387,479
IssuesEvent
2021-11-26 01:50:08
vincetiu8/zombie-game
https://api.github.com/repos/vincetiu8/zombie-game
closed
Add particle hit effects
area/player area/enemies type/feature size/xs priority/low
Add particle effects to let the player know that they have hit a zombie or a zombie has hit a player. This can be number particles which tell how much damage was dealt, blood particles or both
1.0
Add particle hit effects - Add particle effects to let the player know that they have hit a zombie or a zombie has hit a player. This can be number particles which tell how much damage was dealt, blood particles or both
non_test
add particle hit effects add particle effects to let the player know that they have hit a zombie or a zombie has hit a player this can be number particles which tell how much damage was dealt blood particles or both
0
170,694
13,198,368,989
IssuesEvent
2020-08-14 02:14:54
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
[FLAKEY TEST] end-to-end-geth-transfer-test -> celotool -> Transfer tests -> Transfers Frozen > -> light Node > -> when CeloGold is frozen -> when sender is whitelisted -> when sender is removed again from whitelist -> should not add the transaction to the pool
FLAKEY celotool end-to-end-geth-transfer-test
Discovered in commit 4ce1c02c6ad6ca55b90d889a680342fd43b7c221 Attempt No. 1: AssertionError: expected 'Error: Transaction was not mined within 750 seconds, please make sure your transaction was properly sent. Be aware that it might still be mined!' to include 'Returned error: transfers are currently frozen' at /home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:493:16 at Generator.throw (<anonymous>) at rejected (/home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:8:65) at process._tickCallback (internal/process/next_tick.js:68:7) Attempt No. 2: Test Passed!
1.0
[FLAKEY TEST] end-to-end-geth-transfer-test -> celotool -> Transfer tests -> Transfers Frozen > -> light Node > -> when CeloGold is frozen -> when sender is whitelisted -> when sender is removed again from whitelist -> should not add the transaction to the pool - Discovered in commit 4ce1c02c6ad6ca55b90d889a680342fd43b7c221 Attempt No. 1: AssertionError: expected 'Error: Transaction was not mined within 750 seconds, please make sure your transaction was properly sent. Be aware that it might still be mined!' to include 'Returned error: transfers are currently frozen' at /home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:493:16 at Generator.throw (<anonymous>) at rejected (/home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:8:65) at process._tickCallback (internal/process/next_tick.js:68:7) Attempt No. 2: Test Passed!
test
end to end geth transfer test celotool transfer tests transfers frozen light node when celogold is frozen when sender is whitelisted when sender is removed again from whitelist should not add the transaction to the pool discovered in commit attempt no assertionerror expected error transaction was not mined within seconds please make sure your transaction was properly sent be aware that it might still be mined to include returned error transfers are currently frozen at home circleci app packages celotool src tests transfer tests ts at generator throw at rejected home circleci app packages celotool src tests transfer tests ts at process tickcallback internal process next tick js attempt no test passed
1
175,779
13,609,877,886
IssuesEvent
2020-09-23 06:20:12
nrwl/nx
https://api.github.com/repos/nrwl/nx
closed
Improve test performance
scope: testing tools type: feature
Hey there, We've been using this package for quite some time now and it has improved the development experience at our company a lot! However, as our project gets larger and larger, one thing that bothers us is the performance of unit tests. Our project is medium-sized with about 75 libraries and 3 apps (Angular and NestJS) and about 350 test suites. It takes about 12 minutes on my machine (windows, very decent specs) to run all tests using `nx affected:test --all --parallel`. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes. By using `nx affected`, we can run the unit tests for PRs in a few minutes, but for bigger PRs with more files, the time it takes to run the tests increases rapidly. _NOTE: If this is something nobody else is experiencing, feel free to close this issue right away (please comment on how you achieved good test performance)._ After looking into a lot of issues, we're pretty confident that ts-jest and typescript compilation time are the main contributing factors. It takes between 10-20 seconds for each library just to get jest to start running the tests. Once jest is running, the tests are fast. Using the fix from #1299 improved performance quite a bit (around 20-30%), but this still is slow for this amount of tests. kulshekhar/ts-jest#1115 is probably part of the reason. We are still very confident that re-compiling all necessary files for each library individually is not a good approach for unit testing. That's why we decided to try to run all our unit tests together. We altered the global `jest.config.js` and are now able to run all unit tests with the Jest CLI in a single test run. The time it takes to run all those tests is about 60 seconds on my pc and 90-100 seconds on a laptop. I did not yet test the speed in our CI environment. That's a massive speedup (over 10x faster), but could still be improved upon, and that's what this issue is actually about. ~~My suggestion is to alter `nx affected:test` to collect all the projects that have changed, then divide them by their needed jest setup (jest-preset-angular with JSDOM, just JSDOM or Node) and run those tests at once. This could be done by adding an entry to the `testMatch` array in the jest config dynamically for every library and would result in a maximum of three test runs.~~ EDIT: After investigating a bit further, Jest supports a config entry called [`projects`](https://jestjs.io/docs/en/configuration.html#projects-arraystring--projectconfig) since Jest v20. This is a perfect fit for this problem. If the global jest config was altered for every new project generated (just like `tsconfig.json` and `nx.json`), it would be possible to run jest from the root directly through the CLI. This would also allow configuring each project individually as needed through its own `jest.config.js` and to run `ng test <library>` just like now. In addition, I would remove the [hack to alter the jest config in a script](https://github.com/nrwl/nx/blob/5109548710b090bf2fcd68596aed27c24fc45d97/packages/jest/src/builders/jest/jest.impl.ts#L69) and instead change the schematic that is used to generate a new angular library to add the config needed for 'jest-preset-angular' to the libraries' `jest.config.js`. `nx affected:test` would then be even easier to implement then in my previous suggestion. Just replace the `projects` entry in the global jest config by an array of changed project paths. Pros: - Unit tests run **a lot** faster - Jest config is cleaner and more exposed - [Jest CLI](https://jestjs.io/docs/en/cli) can be used to run unit tests - [VSCode-Jest](https://github.com/jest-community/vscode-jest) can be used to run unit tests - The coverage report shows the total coverage Cons: - Parsing failed projects is probably a bit more difficult to implement ## Expected Behavior Be able to run unit tests in 1-2 minutes for small and medium-sized projects and even less for PRs. ## Current Behavior It takes about 12 minutes on my machine (windows, very decent specs) to run all tests. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes. ### Context Please provide any relevant information about your setup: @nrwl/angular : 8.11.0 @nrwl/cli : 8.11.0 @nrwl/cypress : 8.11.0 @nrwl/jest : 8.11.0 @nrwl/linter : 8.11.0 @nrwl/nest : 8.11.0 @nrwl/node : 8.11.0 @nrwl/tao : 8.11.0 @nrwl/workspace : 8.11.0 typescript : 3.4.5 jest-preset-angular: 7.1.1 jest: 24.8.0 ts-jest: 24.3.0
1.0
Improve test performance - Hey there, We've been using this package for quite some time now and it has improved the development experience at our company a lot! However, as our project gets larger and larger, one thing that bothers us is the performance of unit tests. Our project is medium-sized with about 75 libraries and 3 apps (Angular and NestJS) and about 350 test suites. It takes about 12 minutes on my machine (windows, very decent specs) to run all tests using `nx affected:test --all --parallel`. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes. By using `nx affected`, we can run the unit tests for PRs in a few minutes, but for bigger PRs with more files, the time it takes to run the tests increases rapidly. _NOTE: If this is something nobody else is experiencing, feel free to close this issue right away (please comment on how you achieved good test performance)._ After looking into a lot of issues, we're pretty confident that ts-jest and typescript compilation time are the main contributing factors. It takes between 10-20 seconds for each library just to get jest to start running the tests. Once jest is running, the tests are fast. Using the fix from #1299 improved performance quite a bit (around 20-30%), but this still is slow for this amount of tests. kulshekhar/ts-jest#1115 is probably part of the reason. We are still very confident that re-compiling all necessary files for each library individually is not a good approach for unit testing. That's why we decided to try to run all our unit tests together. We altered the global `jest.config.js` and are now able to run all unit tests with the Jest CLI in a single test run. The time it takes to run all those tests is about 60 seconds on my pc and 90-100 seconds on a laptop. I did not yet test the speed in our CI environment. That's a massive speedup (over 10x faster), but could still be improved upon, and that's what this issue is actually about. ~~My suggestion is to alter `nx affected:test` to collect all the projects that have changed, then divide them by their needed jest setup (jest-preset-angular with JSDOM, just JSDOM or Node) and run those tests at once. This could be done by adding an entry to the `testMatch` array in the jest config dynamically for every library and would result in a maximum of three test runs.~~ EDIT: After investigating a bit further, Jest supports a config entry called [`projects`](https://jestjs.io/docs/en/configuration.html#projects-arraystring--projectconfig) since Jest v20. This is a perfect fit for this problem. If the global jest config was altered for every new project generated (just like `tsconfig.json` and `nx.json`), it would be possible to run jest from the root directly through the CLI. This would also allow configuring each project individually as needed through its own `jest.config.js` and to run `ng test <library>` just like now. In addition, I would remove the [hack to alter the jest config in a script](https://github.com/nrwl/nx/blob/5109548710b090bf2fcd68596aed27c24fc45d97/packages/jest/src/builders/jest/jest.impl.ts#L69) and instead change the schematic that is used to generate a new angular library to add the config needed for 'jest-preset-angular' to the libraries' `jest.config.js`. `nx affected:test` would then be even easier to implement then in my previous suggestion. Just replace the `projects` entry in the global jest config by an array of changed project paths. Pros: - Unit tests run **a lot** faster - Jest config is cleaner and more exposed - [Jest CLI](https://jestjs.io/docs/en/cli) can be used to run unit tests - [VSCode-Jest](https://github.com/jest-community/vscode-jest) can be used to run unit tests - The coverage report shows the total coverage Cons: - Parsing failed projects is probably a bit more difficult to implement ## Expected Behavior Be able to run unit tests in 1-2 minutes for small and medium-sized projects and even less for PRs. ## Current Behavior It takes about 12 minutes on my machine (windows, very decent specs) to run all tests. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes. ### Context Please provide any relevant information about your setup: @nrwl/angular : 8.11.0 @nrwl/cli : 8.11.0 @nrwl/cypress : 8.11.0 @nrwl/jest : 8.11.0 @nrwl/linter : 8.11.0 @nrwl/nest : 8.11.0 @nrwl/node : 8.11.0 @nrwl/tao : 8.11.0 @nrwl/workspace : 8.11.0 typescript : 3.4.5 jest-preset-angular: 7.1.1 jest: 24.8.0 ts-jest: 24.3.0
test
improve test performance hey there we ve been using this package for quite some time now and it has improved the development experience at our company a lot however as our project gets larger and larger one thing that bothers us is the performance of unit tests our project is medium sized with about libraries and apps angular and nestjs and about test suites it takes about minutes on my machine windows very decent specs to run all tests using nx affected test all parallel in our ci environment linux and on laptops this can take quite a bit longer around to minutes by using nx affected we can run the unit tests for prs in a few minutes but for bigger prs with more files the time it takes to run the tests increases rapidly note if this is something nobody else is experiencing feel free to close this issue right away please comment on how you achieved good test performance after looking into a lot of issues we re pretty confident that ts jest and typescript compilation time are the main contributing factors it takes between seconds for each library just to get jest to start running the tests once jest is running the tests are fast using the fix from improved performance quite a bit around but this still is slow for this amount of tests kulshekhar ts jest is probably part of the reason we are still very confident that re compiling all necessary files for each library individually is not a good approach for unit testing that s why we decided to try to run all our unit tests together we altered the global jest config js and are now able to run all unit tests with the jest cli in a single test run the time it takes to run all those tests is about seconds on my pc and seconds on a laptop i did not yet test the speed in our ci environment that s a massive speedup over faster but could still be improved upon and that s what this issue is actually about my suggestion is to alter nx affected test to collect all the projects that have changed then divide them by their needed jest setup jest preset angular with jsdom just jsdom or node and run those tests at once this could be done by adding an entry to the testmatch array in the jest config dynamically for every library and would result in a maximum of three test runs edit after investigating a bit further jest supports a config entry called since jest this is a perfect fit for this problem if the global jest config was altered for every new project generated just like tsconfig json and nx json it would be possible to run jest from the root directly through the cli this would also allow configuring each project individually as needed through its own jest config js and to run ng test just like now in addition i would remove the and instead change the schematic that is used to generate a new angular library to add the config needed for jest preset angular to the libraries jest config js nx affected test would then be even easier to implement then in my previous suggestion just replace the projects entry in the global jest config by an array of changed project paths pros unit tests run a lot faster jest config is cleaner and more exposed can be used to run unit tests can be used to run unit tests the coverage report shows the total coverage cons parsing failed projects is probably a bit more difficult to implement expected behavior be able to run unit tests in minutes for small and medium sized projects and even less for prs current behavior it takes about minutes on my machine windows very decent specs to run all tests in our ci environment linux and on laptops this can take quite a bit longer around to minutes context please provide any relevant information about your setup nrwl angular nrwl cli nrwl cypress nrwl jest nrwl linter nrwl nest nrwl node nrwl tao nrwl workspace typescript jest preset angular jest ts jest
1
228,525
18,239,839,152
IssuesEvent
2021-10-01 11:33:32
junit-team/junit5
https://api.github.com/repos/junit-team/junit5
opened
Support text blocks as files in `@CsvSource`
type: enhancement component: Jupiter theme: parameterized tests
## Overview PR #2721 introduced experimental support for _text blocks_ in `@CsvSource`; however, there is room for improvement. A CSV line within a text block cannot currently contain a new line (`\n`), even if it is within a quoted string; whereas, this is supported when using `@CsvSource(value = ...)`. Comments do not make sense in a single string in `@CsvSource(value = ...)`, but they do make sense within `@CsvSource(textBlock = ...)`. ## Deliverables - [ ] Support new lines within "CSV lines" in text blocks - [ ] Support comment lines beginning with `#` in text blocks
1.0
Support text blocks as files in `@CsvSource` - ## Overview PR #2721 introduced experimental support for _text blocks_ in `@CsvSource`; however, there is room for improvement. A CSV line within a text block cannot currently contain a new line (`\n`), even if it is within a quoted string; whereas, this is supported when using `@CsvSource(value = ...)`. Comments do not make sense in a single string in `@CsvSource(value = ...)`, but they do make sense within `@CsvSource(textBlock = ...)`. ## Deliverables - [ ] Support new lines within "CSV lines" in text blocks - [ ] Support comment lines beginning with `#` in text blocks
test
support text blocks as files in csvsource overview pr introduced experimental support for text blocks in csvsource however there is room for improvement a csv line within a text block cannot currently contain a new line n even if it is within a quoted string whereas this is supported when using csvsource value comments do not make sense in a single string in csvsource value but they do make sense within csvsource textblock deliverables support new lines within csv lines in text blocks support comment lines beginning with in text blocks
1
142,193
21,687,146,583
IssuesEvent
2022-05-09 12:24:45
department-of-veterans-affairs/vets-design-system-documentation
https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation
closed
Flesh out Address block Component documentation
vsp-design-system-team va-address-block
## This update is for: - [ ] Content styleguide - [x] Component - [ ] Pattern - [ ] Utility - [ ] Other ## What is the name Address block Component ## What is the nature of this update? - [ ] How to build this component/pattern - [x] When to use this component/pattern - [x] When to use something else - [x] Usage guidance - [x] Accessibility - [ ] Implementation - [ ] Research insights - [ ] Package information ## Additional Context Currently there is no content in the design.va.gov [Address block Component page](https://design.va.gov/components/address-block) other than an embedded Storybook example of the component. We should apply the new Component template to Address block and flesh out this documentation, including links to the Content style guide. We also need to add an example that shows a five-digit zip code, per the advice of Content. Designers have wondered if the full nine-digit zip code is a requirement (it's not) when they see the current Address block example in design.va.gov.
1.0
Flesh out Address block Component documentation - ## This update is for: - [ ] Content styleguide - [x] Component - [ ] Pattern - [ ] Utility - [ ] Other ## What is the name Address block Component ## What is the nature of this update? - [ ] How to build this component/pattern - [x] When to use this component/pattern - [x] When to use something else - [x] Usage guidance - [x] Accessibility - [ ] Implementation - [ ] Research insights - [ ] Package information ## Additional Context Currently there is no content in the design.va.gov [Address block Component page](https://design.va.gov/components/address-block) other than an embedded Storybook example of the component. We should apply the new Component template to Address block and flesh out this documentation, including links to the Content style guide. We also need to add an example that shows a five-digit zip code, per the advice of Content. Designers have wondered if the full nine-digit zip code is a requirement (it's not) when they see the current Address block example in design.va.gov.
non_test
flesh out address block component documentation this update is for content styleguide component pattern utility other what is the name address block component what is the nature of this update how to build this component pattern when to use this component pattern when to use something else usage guidance accessibility implementation research insights package information additional context currently there is no content in the design va gov other than an embedded storybook example of the component we should apply the new component template to address block and flesh out this documentation including links to the content style guide we also need to add an example that shows a five digit zip code per the advice of content designers have wondered if the full nine digit zip code is a requirement it s not when they see the current address block example in design va gov
0
126,097
10,383,582,403
IssuesEvent
2019-09-10 09:57:40
RIOT-OS/RIOT
https://api.github.com/repos/RIOT-OS/RIOT
closed
Testing
Area: CI Area: tests Community: help wanted State: stale
Automated unit tests with hardware in the loop (SAMR21 plugged on CI server?) - Related issues - [ ] #3363 - [ ] #3392 - [ ] #7871 - Related PRs - [ ] #7653 - [x] #7845 - [x] #7906 Automated network functionality tests (e.g. RPL + UDP/PING tests through border router, multi-hop) in IoTLAB dev sites? leverage PiFleet more? - Related issues - [ ] #3252 On-board CI testing in IoT-LAB (as it will provide soon the possibility to add custom nodes) - Related issues - [ ] General CI testing - Related issues - [ ] #2143 - [ ] #5319 - Related PRs - [ ] #7258 - [ ] #7786
1.0
Testing - Automated unit tests with hardware in the loop (SAMR21 plugged on CI server?) - Related issues - [ ] #3363 - [ ] #3392 - [ ] #7871 - Related PRs - [ ] #7653 - [x] #7845 - [x] #7906 Automated network functionality tests (e.g. RPL + UDP/PING tests through border router, multi-hop) in IoTLAB dev sites? leverage PiFleet more? - Related issues - [ ] #3252 On-board CI testing in IoT-LAB (as it will provide soon the possibility to add custom nodes) - Related issues - [ ] General CI testing - Related issues - [ ] #2143 - [ ] #5319 - Related PRs - [ ] #7258 - [ ] #7786
test
testing automated unit tests with hardware in the loop plugged on ci server related issues related prs automated network functionality tests e g rpl udp ping tests through border router multi hop in iotlab dev sites leverage pifleet more related issues on board ci testing in iot lab as it will provide soon the possibility to add custom nodes related issues general ci testing related issues related prs
1
53,725
13,198,209,515
IssuesEvent
2020-08-14 01:42:18
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
GCC truncation warnings with --enable-debug on Fedora 32
Type: Building good first issue
### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Fedora Distribution Version | 32 Linux Kernel | 5.7.14-200.fc32.x86_64 Architecture | x86-64 ZFS Version | master (faa296c73c7ccd535c0874d4f7e8f7c4ea43eea6) <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing I'm seeing truncation warnings when building ZFS using `./configure --enable-debug` on Fedora 32. If I build without `--enable-debug`, I don't see the warnings. ### Describe how to reproduce the problem ``` ./autogen.sh && ./configure --enable-debug && make ``` ### Include any warning/errors/backtraces from the system logs ``` CC os/linux/libzfs_mount_os.lo CC os/linux/libzfs_pool_os.lo CC os/linux/libzfs_sendrecv_os.lo os/linux/libzfs_mount_os.c: In function ‘zfs_selinux_setcontext’: os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4093 [-Werror=format-truncation=] 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~ ...... 268 | append_mntopt(name, context, mntopts, mtabopt, B_TRUE); | ~~~~~~~ os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output 5 or more bytes (assuming 4100) into a destination of size 4096 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ os/linux/libzfs_mount_os.c: In function ‘zfs_adjust_mount_options’: os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4086 [-Werror=format-truncation=] 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~ ...... 297 | append_mntopt(MNTOPT_CONTEXT, prop, | ~~~~ os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output between 12 and 4107 bytes into a destination of size 4096 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ cc1: all warnings being treated as errors make[3]: *** [Makefile:966: os/linux/libzfs_mount_os.lo] Error 1 make[3]: *** Waiting for unfinished jobs.... make[3]: Leaving directory '/home/fedora/zfs/lib/libzfs' make[2]: *** [Makefile:660: all-recursive] Error 1 make[2]: Leaving directory '/home/fedora/zfs/lib' make[1]: *** [Makefile:870: all-recursive] Error 1 make[1]: Leaving directory '/home/fedora/zfs' make: *** [Makefile:733: all] Error 2 ```
1.0
GCC truncation warnings with --enable-debug on Fedora 32 - ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Fedora Distribution Version | 32 Linux Kernel | 5.7.14-200.fc32.x86_64 Architecture | x86-64 ZFS Version | master (faa296c73c7ccd535c0874d4f7e8f7c4ea43eea6) <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing I'm seeing truncation warnings when building ZFS using `./configure --enable-debug` on Fedora 32. If I build without `--enable-debug`, I don't see the warnings. ### Describe how to reproduce the problem ``` ./autogen.sh && ./configure --enable-debug && make ``` ### Include any warning/errors/backtraces from the system logs ``` CC os/linux/libzfs_mount_os.lo CC os/linux/libzfs_pool_os.lo CC os/linux/libzfs_sendrecv_os.lo os/linux/libzfs_mount_os.c: In function ‘zfs_selinux_setcontext’: os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4093 [-Werror=format-truncation=] 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~ ...... 268 | append_mntopt(name, context, mntopts, mtabopt, B_TRUE); | ~~~~~~~ os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output 5 or more bytes (assuming 4100) into a destination of size 4096 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ os/linux/libzfs_mount_os.c: In function ‘zfs_adjust_mount_options’: os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4086 [-Werror=format-truncation=] 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~ ...... 297 | append_mntopt(MNTOPT_CONTEXT, prop, | ~~~~ os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output between 12 and 4107 bytes into a destination of size 4096 250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ cc1: all warnings being treated as errors make[3]: *** [Makefile:966: os/linux/libzfs_mount_os.lo] Error 1 make[3]: *** Waiting for unfinished jobs.... make[3]: Leaving directory '/home/fedora/zfs/lib/libzfs' make[2]: *** [Makefile:660: all-recursive] Error 1 make[2]: Leaving directory '/home/fedora/zfs/lib' make[1]: *** [Makefile:870: all-recursive] Error 1 make[1]: Leaving directory '/home/fedora/zfs' make: *** [Makefile:733: all] Error 2 ```
non_test
gcc truncation warnings with enable debug on fedora system information type version name distribution name fedora distribution version linux kernel architecture zfs version master commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing i m seeing truncation warnings when building zfs using configure enable debug on fedora if i build without enable debug i don t see the warnings describe how to reproduce the problem autogen sh configure enable debug make include any warning errors backtraces from the system logs cc os linux libzfs mount os lo cc os linux libzfs pool os lo cc os linux libzfs sendrecv os lo os linux libzfs mount os c in function ‘zfs selinux setcontext’ os linux libzfs mount os c error ‘ s’ directive output may be truncated writing up to bytes into a region of size snprintf tmp mnt line max quote s s s s name val append mntopt name context mntopts mtabopt b true os linux libzfs mount os c note ‘snprintf’ output or more bytes assuming into a destination of size snprintf tmp mnt line max quote s s s s name val os linux libzfs mount os c in function ‘zfs adjust mount options’ os linux libzfs mount os c error ‘ s’ directive output may be truncated writing up to bytes into a region of size snprintf tmp mnt line max quote s s s s name val append mntopt mntopt context prop os linux libzfs mount os c note ‘snprintf’ output between and bytes into a destination of size snprintf tmp mnt line max quote s s s s name val all warnings being treated as errors make error make waiting for unfinished jobs make leaving directory home fedora zfs lib libzfs make error make leaving directory home fedora zfs lib make error make leaving directory home fedora zfs make error
0
291,127
21,916,024,209
IssuesEvent
2022-05-21 20:54:29
MonoGame/MonoGame
https://api.github.com/repos/MonoGame/MonoGame
opened
Upgrading the samples to 3.8.1
Documentation Help Wanted Good First Issue
Throughout the years, MonoGame has accumulated a bunch of samples. Unfortunately they are all scattered across multiple repositories and are not all ready-to-use for MonoGame 3.8.1 which makes consuming the documentation complicated. A list of those samples is [available here](https://docs.monogame.net/articles/samples.html). We would to unify the samples under the same repository and have the sample documentation page be more straightforward. There are 3 sub-tasks related to this issue. ## 1. Upgrading the official samples to MonoGame 3.8.1 This basically is upgrading [the current samples repository](https://github.com/MonoGame/MonoGame.Samples) to use MonoGame 3.8.1. This should be straightforward. ## 2. Upgrading the former XNA samples and merging them into the main samples repository @SimonDarksideJ has [a giant repository archive of the old XNA 4.0 samples](https://github.com/SimonDarksideJ/XNAGameStudio) (including the ones from old websites now defunct). Some of them are now irrelevant, but it would be interesting to port the most relevant ones to MonoGame 3.8.1 and merge them into [the main samples](https://github.com/MonoGame/MonoGame.Samples) (Platformer and NeonShooter are already ported but not up-to-date with MG 3.8.1). Be mindful of respecting the licenses of each of these samples. We are allowed to use the Microsoft samples as long as we mention ```© Microsoft all rights reserved```. ## 3. Upgrading and merging more samples @CartBlanche has even more [samples](https://github.com/CartBlanche/MonoGame-Samples) archived which we could pick, upgrade, and merge. They seem to be for macOS and some of them are networking samples (which we dropped), so there's some clean up to do. Most have precompiled xnb, so they may be unexploitable. Some seem to come from a book, so we have to be mindful of licenses there too.
1.0
Upgrading the samples to 3.8.1 - Throughout the years, MonoGame has accumulated a bunch of samples. Unfortunately they are all scattered across multiple repositories and are not all ready-to-use for MonoGame 3.8.1 which makes consuming the documentation complicated. A list of those samples is [available here](https://docs.monogame.net/articles/samples.html). We would to unify the samples under the same repository and have the sample documentation page be more straightforward. There are 3 sub-tasks related to this issue. ## 1. Upgrading the official samples to MonoGame 3.8.1 This basically is upgrading [the current samples repository](https://github.com/MonoGame/MonoGame.Samples) to use MonoGame 3.8.1. This should be straightforward. ## 2. Upgrading the former XNA samples and merging them into the main samples repository @SimonDarksideJ has [a giant repository archive of the old XNA 4.0 samples](https://github.com/SimonDarksideJ/XNAGameStudio) (including the ones from old websites now defunct). Some of them are now irrelevant, but it would be interesting to port the most relevant ones to MonoGame 3.8.1 and merge them into [the main samples](https://github.com/MonoGame/MonoGame.Samples) (Platformer and NeonShooter are already ported but not up-to-date with MG 3.8.1). Be mindful of respecting the licenses of each of these samples. We are allowed to use the Microsoft samples as long as we mention ```© Microsoft all rights reserved```. ## 3. Upgrading and merging more samples @CartBlanche has even more [samples](https://github.com/CartBlanche/MonoGame-Samples) archived which we could pick, upgrade, and merge. They seem to be for macOS and some of them are networking samples (which we dropped), so there's some clean up to do. Most have precompiled xnb, so they may be unexploitable. Some seem to come from a book, so we have to be mindful of licenses there too.
non_test
upgrading the samples to throughout the years monogame has accumulated a bunch of samples unfortunately they are all scattered across multiple repositories and are not all ready to use for monogame which makes consuming the documentation complicated a list of those samples is we would to unify the samples under the same repository and have the sample documentation page be more straightforward there are sub tasks related to this issue upgrading the official samples to monogame this basically is upgrading to use monogame this should be straightforward upgrading the former xna samples and merging them into the main samples repository simondarksidej has including the ones from old websites now defunct some of them are now irrelevant but it would be interesting to port the most relevant ones to monogame and merge them into platformer and neonshooter are already ported but not up to date with mg be mindful of respecting the licenses of each of these samples we are allowed to use the microsoft samples as long as we mention © microsoft all rights reserved upgrading and merging more samples cartblanche has even more archived which we could pick upgrade and merge they seem to be for macos and some of them are networking samples which we dropped so there s some clean up to do most have precompiled xnb so they may be unexploitable some seem to come from a book so we have to be mindful of licenses there too
0
182,617
30,874,397,501
IssuesEvent
2023-08-03 13:26:53
elastic/elastic-charts
https://api.github.com/repos/elastic/elastic-charts
opened
[partition] improve pie chart labelling
enhancement :partition design design need
Even if I strongly believe in the goodness of the current implementation of labels in pie chart, I believe we need to make an effort to improve the followings: - find a way to show every label (inside or outside). Right now there are situation where outside labels are not represented. - by default don't rotate the labels inside the slices, if doesn't fit put the label outside - improve alignment of wrapped lines inside the slices (don't make them start from different positions where is not necessary) <img width="488" alt="Screenshot 2023-08-03 at 15 21 54" src="https://github.com/elastic/elastic-charts/assets/1421091/bbe4d658-9dab-4c53-a4fa-2ce038ff63c1">
2.0
[partition] improve pie chart labelling - Even if I strongly believe in the goodness of the current implementation of labels in pie chart, I believe we need to make an effort to improve the followings: - find a way to show every label (inside or outside). Right now there are situation where outside labels are not represented. - by default don't rotate the labels inside the slices, if doesn't fit put the label outside - improve alignment of wrapped lines inside the slices (don't make them start from different positions where is not necessary) <img width="488" alt="Screenshot 2023-08-03 at 15 21 54" src="https://github.com/elastic/elastic-charts/assets/1421091/bbe4d658-9dab-4c53-a4fa-2ce038ff63c1">
non_test
improve pie chart labelling even if i strongly believe in the goodness of the current implementation of labels in pie chart i believe we need to make an effort to improve the followings find a way to show every label inside or outside right now there are situation where outside labels are not represented by default don t rotate the labels inside the slices if doesn t fit put the label outside improve alignment of wrapped lines inside the slices don t make them start from different positions where is not necessary img width alt screenshot at src
0
260,184
22,599,148,800
IssuesEvent
2022-06-29 07:33:11
qaul/qaul.net
https://api.github.com/repos/qaul/qaul.net
closed
Testing Findings, January 14 2022
Discussion Testing Tracking Issue
## General Due to a bug in the time calculation when sending RPC messages to the GUI, libqaul crashed (see logs below). This crash was not transparent to the user, as the UI was still working. This led to confusion. * [x] Fix libqaul bug time calculation bug * libqaul crashed due to time.rs panic (duration became negative) * Polkadot had the same problem: https://github.com/paritytech/polkadot/issues/4096 * [x] find safer way of using time calculations * #295 ## Better Testing Feedback * [x] #291 * [x] Log libqaul within flutter and make it accessible for analyzation in futher tests. * e.g. library quick_log https://pub.dev/packages/quick_log * [x] Create the functionality to send the logs via email to the developers. * [x] Create the functionality to send device soft/hardware information via email to the developers. * [x] #297 * [x] #298 ## UI Improvements * [x] Android: Make System back button working * [x] Mobile: Portrait view: Add gesture sliding functionality between the taps * [x] Feed View: make user icons clickable * [x] User Account View: show connectivity option in view * Display IP addresses via which the user is reachable * [x] #293 * [x] Display round trip time milliseconds on network view correctly. ## Further Development Suggestions - Synchronize feed - Make User Validation Interactive - Automatically exchange further more private information - Explain Idea behind it to the user - Create anti-spam feature for feed messenger ---- ## Logs Libqaul Panic Log ``` flutter: sendRpc send 84 bytes flutter: sendRpc success thread '<unnamed>' panicked at 'overflow when subtracting durations', library/core/src/time.rs:940:31 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace flutter: sendRpc send 84 bytes ERROR libqaul::rpc > "SendError(..)" flutter: sendRpc success ```
1.0
Testing Findings, January 14 2022 - ## General Due to a bug in the time calculation when sending RPC messages to the GUI, libqaul crashed (see logs below). This crash was not transparent to the user, as the UI was still working. This led to confusion. * [x] Fix libqaul bug time calculation bug * libqaul crashed due to time.rs panic (duration became negative) * Polkadot had the same problem: https://github.com/paritytech/polkadot/issues/4096 * [x] find safer way of using time calculations * #295 ## Better Testing Feedback * [x] #291 * [x] Log libqaul within flutter and make it accessible for analyzation in futher tests. * e.g. library quick_log https://pub.dev/packages/quick_log * [x] Create the functionality to send the logs via email to the developers. * [x] Create the functionality to send device soft/hardware information via email to the developers. * [x] #297 * [x] #298 ## UI Improvements * [x] Android: Make System back button working * [x] Mobile: Portrait view: Add gesture sliding functionality between the taps * [x] Feed View: make user icons clickable * [x] User Account View: show connectivity option in view * Display IP addresses via which the user is reachable * [x] #293 * [x] Display round trip time milliseconds on network view correctly. ## Further Development Suggestions - Synchronize feed - Make User Validation Interactive - Automatically exchange further more private information - Explain Idea behind it to the user - Create anti-spam feature for feed messenger ---- ## Logs Libqaul Panic Log ``` flutter: sendRpc send 84 bytes flutter: sendRpc success thread '<unnamed>' panicked at 'overflow when subtracting durations', library/core/src/time.rs:940:31 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace flutter: sendRpc send 84 bytes ERROR libqaul::rpc > "SendError(..)" flutter: sendRpc success ```
test
testing findings january general due to a bug in the time calculation when sending rpc messages to the gui libqaul crashed see logs below this crash was not transparent to the user as the ui was still working this led to confusion fix libqaul bug time calculation bug libqaul crashed due to time rs panic duration became negative polkadot had the same problem find safer way of using time calculations better testing feedback log libqaul within flutter and make it accessible for analyzation in futher tests e g library quick log create the functionality to send the logs via email to the developers create the functionality to send device soft hardware information via email to the developers ui improvements android make system back button working mobile portrait view add gesture sliding functionality between the taps feed view make user icons clickable user account view show connectivity option in view display ip addresses via which the user is reachable display round trip time milliseconds on network view correctly further development suggestions synchronize feed make user validation interactive automatically exchange further more private information explain idea behind it to the user create anti spam feature for feed messenger logs libqaul panic log flutter sendrpc send bytes flutter sendrpc success thread panicked at overflow when subtracting durations library core src time rs note run with rust backtrace environment variable to display a backtrace flutter sendrpc send bytes error libqaul rpc senderror flutter sendrpc success
1
227,487
18,065,158,449
IssuesEvent
2021-09-20 18:13:11
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts - discover feature controls discover feature controls security global discover all privileges allow saving changes to a currently loaded query via the saved query management component
failed-test test-cloud test-failure-flaky v7.15.0 Team:DataDiscovery
**Version: 7.15.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts** **Stack Trace:** ``` Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="saved-query-management-save-changes-button"]) Wait timed out after 10053ms at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17 at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:57:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at Proxy.clickByCssSelector (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/find.ts:360:5) at TestSubjects.click (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:105:5) at SavedQueryManagementComponentService.updateCurrentlyLoadedQuery (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/saved_query_management_component.ts:76:5) at Context.<anonymous> (test/functional/apps/discover/feature_controls/discover_security.ts:134:9) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** - discover feature controls discover feature controls security global discover all privileges allow saving currently loaded query as a copy - discover feature controls discover feature controls security global discover read-only privileges "before all" hook for "shows discover navlink" _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2305/testReport/_
3.0
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts - discover feature controls discover feature controls security global discover all privileges allow saving changes to a currently loaded query via the saved query management component - **Version: 7.15.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts** **Stack Trace:** ``` Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="saved-query-management-save-changes-button"]) Wait timed out after 10053ms at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17 at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:57:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at Proxy.clickByCssSelector (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/find.ts:360:5) at TestSubjects.click (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:105:5) at SavedQueryManagementComponentService.updateCurrentlyLoadedQuery (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/saved_query_management_component.ts:76:5) at Context.<anonymous> (test/functional/apps/discover/feature_controls/discover_security.ts:134:9) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** - discover feature controls discover feature controls security global discover all privileges allow saving currently loaded query as a copy - discover feature controls discover feature controls security global discover read-only privileges "before all" hook for "shows discover navlink" _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2305/testReport/_
test
chrome x pack ui functional x pack test functional apps discover feature controls discover security·ts discover feature controls discover feature controls security global discover all privileges allow saving changes to a currently loaded query via the saved query management component version class chrome x pack ui functional x pack test functional apps discover feature controls discover security·ts stack trace error retry try timeout timeouterror waiting for element to be located by css selector wait timed out after at var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules selenium webdriver lib webdriver js at runmicrotasks at processticksandrejections internal process task queues js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at proxy clickbycssselector var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common find ts at testsubjects click var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common test subjects ts at savedquerymanagementcomponentservice updatecurrentlyloadedquery var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services saved query management component ts at context test functional apps discover feature controls discover security ts at object apply var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules kbn test target node functional test runner lib mocha wrap function js other test failures discover feature controls discover feature controls security global discover all privileges allow saving currently loaded query as a copy discover feature controls discover feature controls security global discover read only privileges before all hook for shows discover navlink test report
1
390,505
26,864,277,577
IssuesEvent
2023-02-03 21:38:12
mindsdb/lightwood
https://api.github.com/repos/mindsdb/lightwood
closed
Improve "ETSMixer" documentation
help wanted good first issue documentation hacktoberfest Community
This mixer currently inherits the docstring from the `sktime` mixer, which translates into duplicated docs :) A simple, short description for the underlying forecasting algorithm would be good enough.
1.0
Improve "ETSMixer" documentation - This mixer currently inherits the docstring from the `sktime` mixer, which translates into duplicated docs :) A simple, short description for the underlying forecasting algorithm would be good enough.
non_test
improve etsmixer documentation this mixer currently inherits the docstring from the sktime mixer which translates into duplicated docs a simple short description for the underlying forecasting algorithm would be good enough
0
134,933
10,948,535,882
IssuesEvent
2019-11-26 09:05:57
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: jepsen/bank-multitable/parts-start-kill-2 failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/02d62674ad2f9ca16183184ea6552691506675f1 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen/bank-multitable/parts-start-kill-2 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1611136&tab=artifacts#/jepsen/bank-multitable/parts-start-kill-2 ``` The test failed on branch=release-19.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20191126-1611136/jepsen/bank-multitable/parts-start-kill-2/run_1 cluster.go:1783,jepsen.go:98,jepsen.go:138,jepsen.go:324,test_runner.go:697: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1611136-1574753130-40-n6cpu4:1-6 -- tar --transform s,^,cockroach/, -c -z -f cockroach.tgz cockroach returned: stderr: stdout: teamcity-1611136-1574753130-40-n6cpu4: tar --transform s,^,cockroa........... 1: 2: 3: 4: exit status 255 5: 6: Error: exit status 255 : exit status 1 ```
2.0
roachtest: jepsen/bank-multitable/parts-start-kill-2 failed - SHA: https://github.com/cockroachdb/cockroach/commits/02d62674ad2f9ca16183184ea6552691506675f1 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen/bank-multitable/parts-start-kill-2 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1611136&tab=artifacts#/jepsen/bank-multitable/parts-start-kill-2 ``` The test failed on branch=release-19.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20191126-1611136/jepsen/bank-multitable/parts-start-kill-2/run_1 cluster.go:1783,jepsen.go:98,jepsen.go:138,jepsen.go:324,test_runner.go:697: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1611136-1574753130-40-n6cpu4:1-6 -- tar --transform s,^,cockroach/, -c -z -f cockroach.tgz cockroach returned: stderr: stdout: teamcity-1611136-1574753130-40-n6cpu4: tar --transform s,^,cockroa........... 1: 2: 3: 4: exit status 255 5: 6: Error: exit status 255 : exit status 1 ```
test
roachtest jepsen bank multitable parts start kill failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests jepsen bank multitable parts start kill pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts jepsen bank multitable parts start kill run cluster go jepsen go jepsen go jepsen go test runner go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity tar transform s cockroach c z f cockroach tgz cockroach returned stderr stdout teamcity tar transform s cockroa exit status error exit status exit status
1
9,621
3,060,347,237
IssuesEvent
2015-08-14 20:08:22
Esri/photo-survey
https://api.github.com/repos/Esri/photo-survey
closed
Photo Survey Application - Testing Survey Submission
help wanted test
Firefox Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Firefox Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted IE 10 & 11 Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted IE 10 & 11 Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Chrome Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Chrome Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Safari Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Safari Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted
1.0
Photo Survey Application - Testing Survey Submission - Firefox Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Firefox Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted IE 10 & 11 Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted IE 10 & 11 Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Chrome Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Chrome Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Safari Submission of Survey (Hosted Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted Safari Submission of Survey (On Premises (AGS) Feature Service) - [x] Test with required questions (proper message displayed on submission) - [x] Answer questions and make sure that the feature service is updated - [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service - [x] Check profile to make survey it is updated to show # of surveys submitted
test
photo survey application testing survey submission firefox submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted firefox submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted ie submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted ie submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted chrome submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted chrome submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted safari submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted safari submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted
1
170,335
13,184,070,480
IssuesEvent
2020-08-12 18:42:04
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
opened
Keyboard Manager requires test for the interaction between the keyboard hook and the UI
Area-Tests Product-Keyboard Shortcut Manager
We should add tests to ensure the correct state variables are set when opening the Remap keys/shortcut window and when clicking the Type button, since the state variables affect the logic of the hook (for eg. it suppresses all key events while the Type window is in the foreground). This should also test that the key remaps are active while the Remap Shortcuts window is in the foreground. The tests will involve testing the read/writes of the `KeyboardManagerUIState` variable (https://github.com/microsoft/PowerToys/blob/master/src/modules/keyboardmanager/common/KeyboardManagerState.h#L45)
1.0
Keyboard Manager requires test for the interaction between the keyboard hook and the UI - We should add tests to ensure the correct state variables are set when opening the Remap keys/shortcut window and when clicking the Type button, since the state variables affect the logic of the hook (for eg. it suppresses all key events while the Type window is in the foreground). This should also test that the key remaps are active while the Remap Shortcuts window is in the foreground. The tests will involve testing the read/writes of the `KeyboardManagerUIState` variable (https://github.com/microsoft/PowerToys/blob/master/src/modules/keyboardmanager/common/KeyboardManagerState.h#L45)
test
keyboard manager requires test for the interaction between the keyboard hook and the ui we should add tests to ensure the correct state variables are set when opening the remap keys shortcut window and when clicking the type button since the state variables affect the logic of the hook for eg it suppresses all key events while the type window is in the foreground this should also test that the key remaps are active while the remap shortcuts window is in the foreground the tests will involve testing the read writes of the keyboardmanageruistate variable
1